The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

The actual Mongo output tool does not allow to specify field type (except for the primary _id field).  The tool just assumes that all fields are string.  Many of our CSV files incorporate string representation of ObjectId (ex: "56df422c08420b523aa00a77"). 

 

When we import those CSV we have to run a an additional script that will convert all the id into ObjectId fields.  Same thing with the date -> Mongo Date.

 

If the tool would allow us to do this, it would save us a billion time across all our ETL processes.

 

Regards,

 

Our development team prefers that we connect to their MongoDB server using a private key through a .pem file, instead of a username/password. Could this option be built into the MongoDB Input?

 

Clustering your data on a sample and then appending clusters is a common theme

especially if you are in customer relations and marketing related divisions...

 Picture2.png

 

When it comes to appending clusters that you have calculated form a 20K sample and then you're going to "score" a few million clients you still need to download the data and use the append cluster...

 

Why don't we have an In-db append cluster instead,

which will quicken the "distance based" scoring that append cluster does

on SQLServer, Oracle or Teradata...

Picture4.png

 

 

 

 Best

 

 

 

Hi All,

 

It would be great if Alteryx 10.5 supports connectivity to SAS server.

 

Regards,
Gaurav

I want to use Alteryx to pull data from a SharePoint List.  This shouldn't be a problem, but I use SharePoint Content Types.  Alteryx won't allow me to import any list that has Content Types enabled; thus rendering the SP list input type not usable. 

My interim workaround is to create a data connection thru excel to the list and then pull the data in that way, but optimally, I would like to pull directly from the list.

 

Content types are a best practice in SharePoint, so any list or library in my site collection contains them.

 

Please update the SharePoint list input to support content types.

 

thank you,

 

Someone else inquired about this but I didn't see an idea entered /

https://community.alteryx.com/t5/Data-Sources/SharePoint-List-Input-Problems-importing-content-types...

 

Hi All,

 

It would be a given wherein IT would have invested effort and time building workflows and other components using some of the tools which became deprecated with the latest versions.

 

It is good to have the deprecated versions still available to make the code backward compatible, but at the same time there should be some option where in a deprecated tool can be promoted to the new tool available without impacting the code.

 

Following are the benefits of this approach -

 

1) IT team can leverage the benefits of the new tool over existing and deprecated tools. For e.g. in my case I am using Salesforce connectors extensibly, I believe in contrast to the existing ones the new ones are using Bulk API and hence are relatively much faster.


2) It will save IT from reconfiguring/recoding the existing code and would save them considerable time.

 

3) As the tool keeps forward moving in its journey, it might help and make more sense to actually remove some of the deprecated tool versions (i.e. I believe it would not be the plan to have say 5 working set of Salesforce Input connectors - including deprecated ones). With this approach in place I think IT would be comfortable with removal of deprecated connectors, as they would have the promote option without impacting exsiting code - so it would ideally take minimal change time.

 

In addition, if it is felt that with new tools some configurations has changed (should ideally be minor), those can be published and as part of
promotions IT can be given the option to configure it.

 

Thanks,

Rohit Bajaj

I find that when I'm using Alteryx, I'm constantly renaming the tool connectors. Here's my logic, most of the time:

I have something like a Join and 3 browses.

  1. I name the L join something like "L: on product/location join"
  2. I then copy that descriptor, and past it in the Annotation field
  3. I then copy that descriptor, select the wired connector, and paste that in the connection configuration

MY VISION:

Have a setting where I could select the following options:

  • Automatically annotate based on tool rename
  • Automatically rename incoming connector based on tool rename

alteryx auto rename example.PNGTHE GUTS/REQUEST:

If I rename a tool, and "Automatically annotate based on tool rename" is enacted, it will insert that renaming at the top of the annotation field. If there is already data in that field, it will be shifted down. If I rename a tool and "Automatically rename incoming connector..." is on, then the connection coming into it gets [name string]+' connection' put into its name field. I included a picture of the end game of my request.

 

Thanks for your ear!

 

When users who have no idea what the Field API Name is try to pull data from Salesforce it can be problematic. A simple solution would be to add the Field Label to the Query Window to allow users to pick the fields based on API NAME or FIELD LABEL.

 When pulling data from Oracle and pushing to Salesforce, there are many times where we have an ID field in Oracle and a field containing this ID in Salesforce in what is called an external ID field. Allowing us to match against those external ID fields would save us a lot of time and prevent us from having to do a query on the entire object in Salesforce to pull out the ID of the records we need.

Sometimes in a crowded workflow, connector lines bunch up and align across the title bar of a tool container.  This blocks my view of the title, but also makes it hard to 'grab' the tool container and move it.

 

Could Alteryx divert lines around tool containers that they don't connect into, or make tool containers 'grab-able' at locations other than the title bar?

 

Image demonstrating connector line overlapImage demonstrating connector line overlap

Hi,

 

When you create a data model in excel you can create measures (aka KPI). These is then something you can then use when you pivot the data and measure would dynamically be updated as you segment the data in your pivot table.

 

By example, let's say you have a field with the customer name and a field with the revenue, you could create a measure that will calculate the average revenue per customer (sum of revenue / disctintcount of customer)

Now if you have a 3rd field in your data that inidcates your region, the measure would allow to see the average revenue by customer and by region (but the measure formula would remain the same and wouldn't refer to the region field at all)

 

Excel integrates well with PowerBI and currently these measures flow into PowerBI.

 

While we have a "Publish to PowerBI" in Alteryx I haven't seen any way to create such measures and export them to PowerBI.

 

Hence I still need to load to Excel to create these measures before I can publish to PowerBI, it'd be great to avoid that intermediate tool.

 

Thanks

 

Tibo

 

We don't have Server.  Sometimes it's easy to share a workflow the old fashioned way - just email a copy of it or drop it in a shared folder somewhere.  When doing that, if the target user doesn't have a given alias on their machine, they'll have issues getting the workflow to run.

 

So, it would be helpful if saving a workflow could save the aliases along with the actual connection information.  Likewise, it would then be nice if someone opening the workflow could add the aliases found therein to their own list of aliases.

 

Granted, there may be difficulties - this is great for connections using integrated authentication, but not so much for userid/password connections. Perhaps (if implemented) it could be limited along these lines.

 

I have run into an issue where the progress does not show the proper number of records after certain pieces in my workflow. It was explained to me that this is because there is only a certain amount of "cashed" data and therefore the number is basd off of that. If I put a browse in I can see the data properly.

 

For my team and me, this is actually a great inconvenience. We have grown to rely on the counts that appear after each tool. The point of the "show progress" is so that I do not have to insert a browse after everything I do so that it takes up less space on my computer. I would like to see the actual number appear again. I don't see why this changed in the first place.

Improve HIVE connector and make writable data available

 

Regards,

Cristian.

 

Hive

 

Type of Support:Read-only
Supported Versions:0.7.1 and later
Client Versions:--
Connection Type:ODBC
Driver Details:The ODBC driver can be downloaded here.

Read-only support to Hive Server 1 and Hive Server 2 is available.

Hello,

I think it would be extremely useful to have a switch connector available in Alteryx. What I mean by a switch connector is a connecting line with an on/off state that will block the data stream through it when off. Something like below:

This image is not available because: You don’t have the privileges to see it, or it has been removed from the system
Switch Connector in an "Off" state

This would be extremely useful when you only want data to flow down some of the paths. In the example above, I might turn the switch connector to off because I want to see the Summarize results without outputting to a document.

The current methods for having a path/set of tools present but unused are insufficient for my needs. The two methods I and Alteryx support were able to find were:

1. Deleting the connecting line - This works, but throws up errors. Even though this is functional, it looks bad when I need to present my Alteryx module and there are errors.
2. Putting the tools in a disabled tool container - I cannot see the tools when the container is disabled. I want to be able to see my tool set-up even when I am not using it.

This is inspired by the use of switches in electrical circuit design, such as:
This image is not available because: You don’t have the privileges to see it, or it has been removed from the system
Please comment if you also think this would be useful, or if you have ideas for ways to improve it further. Thank you!
0 Likes

Hi Team,

 

With Sharepoint Tool 2.3.0 , We are unable to connect Sharepoint Lists with service Principal Authentication as it requires SharePoint - Application permission - Sites.Read.All and Sites.ReadWrite.All in Microsoft Azure App. However, as those permissions will gets access to all sites in respective Organization community, it is impossible for any company to provide as it leaks data security. Kindly provide any alternative or change in permsiions for Sharepoint Connectivity with thumbprint in Alteryx.

 

Regerence Case with Alteryx Support : Case #00619824

 

Thanks & Regards

Vamsi Krishna

0 Likes

HI, 

Not sure if this Idea was already posted (I was not able to find an answer), but let me try to explain.

 

When I am using Mongo DB Input tool to query AlteryxService Mongo DB (in order to identify issues on the Gallery) I have to extract all data from Collection AS_Result.

The problem is that here we have huge amount of data and extracting and then parsing _ServiceData_ (blob) consume time and system resources.

 

This solution I am proposing is to add Sorting option to Mongo input tool. Simple choice ASC or DESC order. 

Sebastian_Chaieb2_0-1679473890913.png

 

Thanks to that I can extract in example last 200 records and do my investigation instead of extracting everything 

 

In addition it will be much easier to estimate daily workload and extract (via scheduler) only this amount of data we need to analyze every day ad load results to external BD.

 

Thanks,

Sebastian 

0 Likes

When you import a csv file,  I sometimes use a "TAB" as delimiter. In section 5 Delimiters I want that as an option.

I have learned that it is possible to wright "\t" but a normal choice would bed nice.

0 Likes

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/BigQuery-Input-Error/td-p/440641

 

The BigQuery Input Tool utilizes the TableData.List JSON API method to pull data from Google Big Query. Per Google here:

 

  • You cannot use the TableDataList JSON API method to retrieve data from a view. For more information, see Tabledata: list.

 

This is not a current supported functionality of the tool. You can post this on our product ideas page to see if it can be implemented in future product. For now, I would recommend pulling the data from the original table itself in BigQuery. 

 

I need to be able query tables and views.  Not sure I know how to use tableDataList JSON API.

0 Likes

The current Azure Data Lake tool appears to lack the ability to connect if fine grained access is enabled within ADLS using Access Control Lists (ACL)

It does work when Role Based Access Control (RBAC) is used by ACL provides more fine grained control to the enviroment

 

For example using any of the current auth types: End-User Basic, End-User (advanced) or Service-to-Service if the user has RBAC to ADLS the connector would work

In that scenario though the user would be granted to an entire container which isn't ideal

  • azureStorageAccount/Container/Directory
    • Example: azureStorageAccount/Bronze/selfService1 or azureStorageAccount/Bronze/selfService2
    • In RBAC the user is granted to the container level and everything below so you cannot set different permissions on selfService 1 or selfService2 which may have different use cases

The ideal authentication would be to the directory level to best control and enable self service data analytic teams to use Alteryx

  • In this access pattern the user would only be granted to the directory level (e.g. selfService1 or selfService2 from above)

The existing tool appears to be limited where if don't have access at the container level but only at the directory level then the tool cannot complete the authentication request.  This would require the input for the tool to be able to select a container (aka file system name) from the drop down that included the container+ the directory

  • Screenshot example A below shows how the file system name would need to be input
  • Screenshot example B below shows what happens if you have ACL access to ADLS at the directory level and not at the container level

 

Access control model for Azure Data Lake Storage Gen2 | Microsoft Docs

 

Example A

AStasi_2-1652972198944.png

 

Example B

AStasi_3-1652972437620.png

 

 

 

 

Top Liked Authors