Community Spring Cleaning week is here! Join your fellow Maveryx in digging through your old posts and marking comments on them as solved. Learn more here!
The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

I have a problem when transferring records between different O365 Sharepoint Sites.  It seems that Alteryx cannot maintain 2 separate connections at the same time.  I can transfer fine if I read from one site to a temp file and then, in another workflow, read from the file and write to the second site.

 

I can work around the problem using Block until Done, but there are some situations where I need to be able to compare between lists in 2 different sites and write back to one or both depending on the results.  it would be much more convenient to be have multiple connections open simultaneously.  I'm aware that Alteryx uses the SharePoint API to move information around.  This API does allow multiple connections.  I'm not familiar with the internals of how Alteryx accesses the API, perhaps the OAuth token is shared through out the workflow process,  but this should be posssible

 

Thanks for considering this

 

Dan

   

 

I have used the SharePoint List Connectors with our SharePoint 2010 (on prem) instance for some time now. It works great and has become invaluable. Unfortunately, I have been told that these connectors do not work with a Cloud instance or an instance that is not on prem. We need this capability since Microsoft is pushing coporations to move to the cloud instance and there is talk that on-prem may not be available past the 2016 version that is coming soon. Many companies including mine have either completed or are close to completing a full migration which has rendered the current SharePoint Connectors useless. While this is the most important part, another piece that is missing is a SharePoint Document Library connector (similar to the Amazon S3 Download/Upload). Currently I must use the UNC path to my SharePoint folders and an easier more reliable way to save files out to OneDrive and SharePoint Online would be very beneficial.

Hyperion Smartview Connect

Goal: My goals is to pull Marketo Lead Data via the Marketo Input SOAP API Connector and have specificied only 3 columns to pull for 1 day. 

Issue: It appears that the Marketo Input SOAP API Connector needs to go through all 500 + columns of lead object data before it filters out the specific request I indicated.  What this means is that to process 1 day of lead data for 3 columns of data, it takes 45 minutes plus and not ideal.

Client Services indicated that there is a limitation with the Marketo SOAP API tool and it would be best to utilize the Download tool and build a custom connection to the CRM System (Microsoft Dynamics) to get a predefined list of Lead IDs and then use that as a filter via the Marketo Input SOAP API connector. 

Request:
1.  Add Microsoft Dynamics Connector
2.  Try to innovate a better Marketo Service Connector to just grab the defined fields as opposed to go through the entire processing for all records to only filter back down to. 

Eric  

Service Now dashboards really don't give me as much flexibility as I need, but can't help thinking an Alteryx connector coupled with Qlik Sense dashboards would be fantastic. SNOW seem to be supportive of Tableau - have you guys got plans to get in on the action?

Adobe analytics api current version is 2.0 but the connector still uses 1.4. This works for log in access but does not for automated access since the end points are changed and 2.0 uses a new key structure

A question has been coming up from several users at my workplace about allowing a column description to display in the Visual Query Builder instead of or along with the column name. 

 

The column names in our database are based on an older naming convention, and sometimes the names aren't that easy to understand. We do see that (if a column does have a column description in metadata) it shows when hovering over the particular column; however, the consensus is that we'd like to reverse this and have the column description displayed with the column name shown on hover. 

 

It would be a huge increase to efficiency and workflow development if this could be implemented.

 

Alternative data sources are key for enriching data. one source is social media.I believe Alteryx lacks in social media analytics.

 

  • I would like to propose a YouTube connector...
  • Crawl comments and video definitions (useful for text analytics)
  • Impression and likes/dislikes... (time-series data)

 

If you are into #media, #advertising, #marketing analytics, #influencer analytics please support the idea by seconding...

 

There is an API called YouTube DATA API

https://developers.google.com/youtube/v3/getting-started

Picture1.png

 

 

 

Shopify is a popular software for e-commerce. With Shopify Connector in Alteryx, it will be very helpful for Alteryx users to integrate with Shopify. 

There is no way to natively connect to Microsoft Dynamics 365 F&O. We also cannot use the download tool to connect to the API as it requires an OAuth2 Bearer Token generation step. Documentation can be found here: https://docs.microsoft.com/en-us/dynamics-nav/api-reference/v1.0/

 

I would request a tool similar to the Salesforce Input tool to be created.

Suggested process:

  • Configuration Window to log into a Microsoft Dynamics 365 F&O account
  • Configuration Window to select the following
    • 1 or more companies ("all" should also be an option)
    • The specific "table" or "api call" you want to query - for example the "Sales Invoice" Data
    • Date Range filter (both from + to date combination as well as some generic "this month", "previous month", "this quarter" kind of selections)

I would also request that this tool be developed so that it can run on Server and Analytics Hub so that we can run end to end automation.

Currently we are limited to chossing one of two layout direction options, vertical or horizontal.  Why not make the direction assignable at the tool icon instead of as a module level control.  I could right click the tool and have layout direction as an option which would activate a visual handle which could either allow infinite rotation control or rotation control in 45 degree increments.  You can use Viso as an example of rotational control for a shape.  In Visio the shape rotates, in our case since we are really looking to change the flow direction the icon could remain in the same orientation as it does now but the conenctor point(s) would rotate around the compass in say 45% increments base on the drag of the rotation handle that appears

Google Big Query can contain so-called complex queries, this means that a table "cell" can host an array. These arrays are currently not supported in the SIMBA ODBC driver for Google Big Query (GBQ), ie. you can't download the full table with arrays.

 

All Google Analytics Premium exports to GBQ contain these arrays, so as an example, please have a look at the dataset provided via this link: 

 

 

https://support.google.com/analytics/answer/3416091?hl=en

 

The way to be able to query these tables is via a specific GBQ SQL query, for instance:

 

Select * From flatten(flatten(flatten(testD113.ga_sessions,customDimensions),hits.customVariables),hits.customDimensions).

 

This is not currently supported from within the Input tool with the Simba ODBC driver for GBQ, which I would like to suggest as a product idea.

 

Thanks,

 

Hans

 

 

It is great to see the ability to stage data for bulk loading into Databricks in s3 and ADLS.  Previously this only appeared to allow staging in Databricks DBFS. 

 

However the current connector included in Designer 2022.1 has a key gap in functionality with ADLS Gen 2 

 

The only authentication method provided to the ADLS storage is though a shared key. 

Shared keys provide access that is

  1. Too broad and allows access to the entire storage
  2. Limited or no auditability and traceability to who is using it 

We do not provide users the shared key for the ADLS storage, thus our users cannot take advantage of this new feature. 

 

The preferred method of authentication to ADLS would be

  1. RBAC - Role Based Access Control
  2. ACL - Access Control lists

Either of these options can be provided though a service principal with a tenant id, client id and client secret as inputs to the bulk load tool

 

This request would specifically be to allow the ACL authentication. ACL would help empower our our self service data analysts and data scientists who could have access to a specific container. 

 

For example

 

storageAccount/Container/directory

 

The ACL access in this tool would allow the Alteryx tool to follow the same access patterns where fine grained access was provided at the directory level and not at the storage account or container level.  This would provide self service analysts and data scientists to use Alteryx as they need within their directory without providing higher level access.  

 

Access control model for Azure Data Lake Storage Gen2 | Microsoft Docs

 

 

AStasi_0-1652966213867.png

 

In alterxy BI teams, we often need to hit the same data-store - however if shared connections are not adopted in the server environment, there's currently no easy way to share the connections among the team (none that I know of)

 

Would be great if there were a "share" button on the data connections tab in the designer, so that we could share connections with a team, or export known good connections to a file that can be kept under config-management / version control.

 

Thank you

Sean

It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive.  For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox.  We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility. 

In the Python 3.6/3.8 versions of SF Input Tool, the business name of an object is returned (e.g., Quote).  In the now-deprecated Macro-based version of SF Input, the technical name was returned (e.g., Quote, Quote__c, SBQQ_Quote__c).

 

With the Python Input tools, there are multiple occurrences of "Quote" to select from with the SF Input tool.  This is confusing and leads to "guessing" which object is the right one.

 

See attached screenshots.

 

My proposal is to add an option to the SF Input tool to allow the workflow developer to choose whether technical or business names should be returned.

Would be great to have a dynamic input tool that is a SFDC connector to update SOQL queries based on the fly. 

The current SharePoint API pull tool does not support the pull of managed metadata columns. It would be great if Alteryx would update the SharePoint List tools to be able to read in managed metadata columns.  

Wild card read option (Sales_*.csv ) works for reading files of the same structure available on a standard windows network/local fileshare. It would be good to have the same feature for reading from HDFS via the HDFS connection option using the input tool.

 

 

 

Regards,

Sandeep.

 

JIRA is used very widely in industry as a defect tracker, issue management system, in many cases even as a super-simple-workflow platform.

 

Could you please consider adding a JIRA connector to Alteryx so that teams can connect to a JIRA instance directly via the API?

 

Many thanks

S

Top Liked Authors