community
cancel
Showing results for 
Search instead for 
Did you mean: 

Alteryx Designer Ideas

Share your Designer product ideas - we're listening!

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the like in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

My specific use case relates to writing to AWS but am sure there are many other use cases for federated user session token support.

 

Specifically, using the S3 Upload tool or Athena Bulk Write (via SIMBA and Athena ODBC), the configuration works when using a IAM user, access key, and secret access key but when using a federated user via Okta there is no option to enter the session token and authentication fails.

Alteryx desktop should support federated users' session tokens.

Please update the Publish to Tableau Server connector tool to support Tableau's Ask Data feature. The data source must be recognized as an extract on Tableau Server in order for the Ask Data feature to work. Currently, all data source published using version 2.0 of the connector tool are recognized as a live data source. The work around is cumbersome and requires multiple copies of data sources to be created and managed.

  • Connectors

In Alteryx enable connections to Oracle Databases that are configured to use External Authentication.

This should allow Alteryx workflows to connect to Oracle databases using different authentication mechanisms, e.g. Kerberos.

 

Please see this discussion for some analysis on what would be required in Alteryx to support Oracle Database External Authentication:

 

Essentially this would involve Alteryx allowing users to specify that a connection to an Oracle database will utilize external authentication.

Then when connecting to an Oracle database with external authentication, Alteryx would pass the relevant parameter to Oracle to indicate external authentication is required (and Alteryx would not pass user name and password info). Then authentication with the Oracle database would be controlled by the external authentication configuration on the computer running Alteryx.

 

For more information on Oracle Database External Authentication see:

  • Connectors

It would be wonderful for Alteryx to be able to connect to and query OData feeds natively, rather than using a 3rd-party driver or custom macro.   

 

OData querying is supported by quite a few familiar products, including Excel and PowerBISSIS/SSRS, FME SafeTableau, and many others. And the protocol is used to publish feeds from Microsoft Dynamics and Sharepoint, as well as many of the 10,000 publically available government datasets with API's (esp. those hosted by Socrata)   

 

I didn't see it as in the Idea section, but questions and workarounds have been discussed in the community a few times (11/15, 3/18, 4/18), and suggestions seem to be just to buy the $400-600 ODBC driver from CDATA (or ZappySys), or I could use a VBA script in Excel trigger a refresh, or create my own Alteryx connector macro (great series btw, though most was beyond my understanding!) 

   

While not opposed paying, kludging, or learning to program, they're just one more thing to build/buy, install, maintain, and break at the most inconvenient time 🙂

 

Thanks,
Chadd

 

OData Overview:

OData (Open Data Protocol) is an ISO/IEC approvedOASIS standard that defines a set of best practices for building and consuming RESTful APIs. OData helps you focus on your business logic while building RESTful APIs without having to worry about the various approaches to define request and response headers, status codes, HTTP methods, URL conventions, media types, payload formats, query options, etc. OData also provides guidance for tracking changes, defining functions/actions for reusable procedures, and sending asynchronous/batch requests.  OData RESTful APIs are easy to consume. The OData metadata, a machine-readable description of the data model of the APIs, enables the creation of powerful generic client proxies and tools.

More info at at http://odata.org

Hi All,

 

Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?

 

For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.

 

Best,

devKev

I have recently added an Azure data lake v2.   The Azure input/output connectors do not work with this version of the Azure data lake.

 

It appears that Alteryx adds ".azuredatalakestore.net" to the file path.   This works for V1, but not needed for V2

 

any plans to configure a connector for Azure data lake v2?

  • Connectors

I know that incoming and outgoing connections can be wired and wireless, and that they will highlight when one clicks on a tool. However, it would be very useful to be able to highlight a particular connector in a particular colour (selected from a palette, perhaps, from the drop down window, or from the configuration). This would be especially useful when there are many connectors originating from a single tool.

Thanks

I have not found this function or a workaround, only as "recent connections" which normally, are not saved on Virtual Machines.

This would save the time it takes to find the path/folder where Calgary DBs are saved.

 

if this has been proposed or fixed already, please delete this idea! 

Hi,

A lot of companies now are deploying on both AWS and Microsoft Azure.

Alteryx supports AWS S3 object storage out of the box, it would be important to support Microsoft Azure blob as part of the native Alteryx product as well. 

Cheers,

Adrian

it would be great to have an Azure SQL DB or/and Azure SQL DW Connector. In the moment it is only possible to connect with the Standard SQL Db connector. To be successful a username and password is always required.

It would be perfect, if you can use your Azure AD User and Password and connect against the relevant Database within the Azure environment. 

  • Connectors

For those of us obsession with the look and feel of our workflows, it would be great to have the option to modify the paths of tool connections, as shown in the image below. 

 

Alteryx Idea.png

 

 

Very nice to be able to extend Alteryx with R programs or CMD execution.  Please, please, please add a connection to Python!
  • Connectors

Create new connector to pull Salesforce Reports

 

We are a large company with tens of thousands employees using Salesforce on a daily basis. Over the years, we have worked with Salesforce to make many customizations and create many reports to provide data for various reporting needs. However, we have increasingly found it inefficient and prone to error to download the reports manually. We have many teams using the Salesforce reports as a base to create additional business insights.

 

Alteryx is a great tool to manage data ETL and workflows, but it does not support pulling data from Salesforce reports directly. Instead, it only offers connectors to pull data from base Salesforce objects. The data from Salesforce objects such as tables can be useful, but do not necessarily offer the logical view of Salesforce reports, and may require a lot of efforts to reconcile the data consistency against the reports our users are used to. Sometimes, it may be impossible to repeat producing the same data from Salesforce tables as those from Salesforce reports. That in turn would cause a lot of efforts spent by the reporting teams, their audience, and users of the Salesforce reports to match things up.

 

Salesforce does not have any out-of-box solution to schedule downloading the reports. At our request, their support team did some research and have not found a good 3rd-party solution in the Salesforce App Exchange ecosystem that supports this need.

 

I strongly believe this is a great opportunity for Alteryx. Salesforce already has an API that allows for building custom applications to pull Salesforce reports. However, most Salesforce users are more business oriented and do not necessarily have the appetite to engage with their IT staff or external resources provide to develop such apps and bear the burden to main them.

 

I have attached the Salesforce Reports and Dashboards API Developer Guide for your reference.

 

Sincerely,

Vincent Wang

Right now the PublishToPowerBI connector only publishes to "My Workspace."  I manage datasets that feed reports for multiple workspaces, some of which are not necessarily personal workspaces (so there is no login associated). A drop-down that lets you select which workspace, that you are a member of, would be fantastic!

 

The workaround right now is to ETL in Alteryx then save the dataset out to OneDrive. You can then "Publish" the Excel sheet to Power BI natively, and the data refreshes once an hour. This works for some data, but we have use cases that need refresh rates much higher than that. Plus publishing directly to Power BI would be ideal.

With more and more enterprises moving to cloud infrastructures and Azure being one of the most used one, there should be support for its authentication service Azure Active Directory (AAD).

 

Currently if you are using cloud services like Azure SQL Servers the only way to connect is with SQL login, which in a corporate environment is insecure and administrative overhead to manage.

 

The only work around I found so far is creating an ODBC 17 connection that supports AAD authentication and connect to it in Alteryx.

 

Please see the post below covering that topic:

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Alteryx-Designer-How-to-Connect-to-Azu...

Please add Parquet data format (https://parquet.apache.org/) as read-write option for Alteryx.

 

Apache Parquet is a columnar storage format available to any project in the Hadoop ecosystem, regardless of the choice of data processing framework, data model or programming language.

 

Thank you.

 

Regards,

Cristian.

The Data Sources page currently lists all the different data sources that Alteryx supports - however for an administrator it's almost impossible to ensure that their designer users have the drivers for these, or are on the right version.

 

As an early step - can we add 1 more field to this list which points to the downloader for the driver where applicable

https://help.alteryx.com/current/DataSources/SupportedDataSources.htm

 

The next step is to create a bundled driver pack that Admins can download once to cover all the relevant drivers needed for the data sources provided & supported.

 

cc: @revathi @AshwiniChezhiyan @LizaNemchynova

 

  • Connectors

I installed version 4.0.0 of the Salesforce Input tool since the old tool is deprecated. When I open a saved workflow that has Salesforce inputs, the inputs display an immediate error, for example "Error: Salesforce Input (95): 401 Client Error: Unauthorized for url: https://servings.my.salesforce.com/services/data/v42.0/sobjects/Task/describe/". The workflow will not run until I click into each Salesforce input and then click outside of it again to make it refresh, which takes a few seconds each time. I'm guessing this has to do with Alteryx saving an expired OAuth access token and trying to pull updated metadata.

It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive.  For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox.  We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility. 

Screen Shots.JPG

I’m writing about a short-coming I see in the Publish to Tableau Server Tool v2.0 (PTTS).  I work in a development environment where we use different Tableau servers (i.e. development, test, production) to support product development.  One of the shortcomings of PTTS is that once the Tableau server information you are connecting to is entered, validated, and the tool is configured, you can no longer ‘see’ which Tableau server/site the tool is publishing to.  I think this piece of information is quite important.  I know I can always us the “Disconnect” button in the tool and re-enter the information so I know which server it is pointing to, but this defeats the purpose of entering that information in the tool in the first place.

 

Please consider an enhancement to the tool so we developers know at a glance where (server/site) the tool is publishing.  Project and Data source names are helpful, but in a development cycle, all Tableau servers may have the same Project and Data source names across all environments.

 

I've attached examples of the tool options when being defined and once the tool is configured – notice server URL and Site are aren’t displayed in the tool once it has been configured.

 

Top Liked Authors