This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
My specific use case relates to writing to AWS but am sure there are many other use cases for federated user session token support.
Specifically, using the S3 Upload tool or Athena Bulk Write (via SIMBA and Athena ODBC), the configuration works when using a IAM user, access key, and secret access key but when using a federated user via Okta there is no option to enter the session token and authentication fails.
Alteryx desktop should support federated users' session tokens.
Please update the Publish to Tableau Server connector tool to support Tableau's Ask Data feature. The data source must be recognized as an extract on Tableau Server in order for the Ask Data feature to work. Currently, all data source published using version 2.0 of the connector tool are recognized as a live data source. The work around is cumbersome and requires multiple copies of data sources to be created and managed.
Essentially this would involve Alteryx allowing users to specify that a connection to an Oracle database will utilize external authentication.
Then when connecting to an Oracle database with external authentication, Alteryx would pass the relevant parameter to Oracle to indicate external authentication is required (and Alteryx would not pass user name and password info). Then authentication with the Oracle database would be controlled by the external authentication configuration on the computer running Alteryx.
For more information on Oracle Database External Authentication see:
I didn't see it as in the Idea section, but questions and workarounds have been discussed in the community a few times (11/15, 3/18, 4/18), and suggestions seem to be just to buy the $400-600 ODBC driver from CDATA (or ZappySys), or I could use a VBA script in Excel trigger a refresh, or create my own Alteryx connector macro (great series btw, though most was beyond my understanding!)
While not opposed paying, kludging, or learning to program, they're just one more thing to build/buy, install, maintain, and break at the most inconvenient time 🙂
OData (Open Data Protocol) is an ISO/IEC approved, OASIS standard that defines a set of best practices for building and consuming RESTful APIs. OData helps you focus on your business logic while building RESTful APIs without having to worry about the various approaches to define request and response headers, status codes, HTTP methods, URL conventions, media types, payload formats, query options, etc. OData also provides guidance for tracking changes, defining functions/actions for reusable procedures, and sending asynchronous/batch requests. OData RESTful APIs are easy to consume. The OData metadata, a machine-readable description of the data model of the APIs, enables the creation of powerful generic client proxies and tools.
Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?
For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.
I know that incoming and outgoing connections can be wired and wireless, and that they will highlight when one clicks on a tool. However, it would be very useful to be able to highlight a particular connector in a particular colour (selected from a palette, perhaps, from the drop down window, or from the configuration). This would be especially useful when there are many connectors originating from a single tool.
it would be great to have an Azure SQL DB or/and Azure SQL DW Connector. In the moment it is only possible to connect with the Standard SQL Db connector. To be successful a username and password is always required.
It would be perfect, if you can use your Azure AD User and Password and connect against the relevant Database within the Azure environment.
We are a large company with tens of thousands employees using Salesforce on a daily basis. Over the years, we have worked with Salesforce to make many customizations and create many reports to provide data for various reporting needs. However, we have increasingly found it inefficient and prone to error to download the reports manually. We have many teams using the Salesforce reports as a base to create additional business insights.
Alteryx is a great tool to manage data ETL and workflows, but it does not support pulling data from Salesforce reports directly. Instead, it only offers connectors to pull data from base Salesforce objects. The data from Salesforce objects such as tables can be useful, but do not necessarily offer the logical view of Salesforce reports, and may require a lot of efforts to reconcile the data consistency against the reports our users are used to. Sometimes, it may be impossible to repeat producing the same data from Salesforce tables as those from Salesforce reports. That in turn would cause a lot of efforts spent by the reporting teams, their audience, and users of the Salesforce reports to match things up.
Salesforce does not have any out-of-box solution to schedule downloading the reports. At our request, their support team did some research and have not found a good 3rd-party solution in the Salesforce App Exchange ecosystem that supports this need.
I strongly believe this is a great opportunity for Alteryx. Salesforce already has an API that allows for building custom applications to pull Salesforce reports. However, most Salesforce users are more business oriented and do not necessarily have the appetite to engage with their IT staff or external resources provide to develop such apps and bear the burden to main them.
I have attached the Salesforce Reports and Dashboards API Developer Guide for your reference.
Right now the PublishToPowerBI connector only publishes to "My Workspace." I manage datasets that feed reports for multiple workspaces, some of which are not necessarily personal workspaces (so there is no login associated). A drop-down that lets you select which workspace, that you are a member of, would be fantastic!
The workaround right now is to ETL in Alteryx then save the dataset out to OneDrive. You can then "Publish" the Excel sheet to Power BI natively, and the data refreshes once an hour. This works for some data, but we have use cases that need refresh rates much higher than that. Plus publishing directly to Power BI would be ideal.
The Data Sources page currently lists all the different data sources that Alteryx supports - however for an administrator it's almost impossible to ensure that their designer users have the drivers for these, or are on the right version.
As an early step - can we add 1 more field to this list which points to the downloader for the driver where applicable
I installed version 4.0.0 of the Salesforce Input tool since the old tool is deprecated. When I open a saved workflow that has Salesforce inputs, the inputs display an immediate error, for example "Error: Salesforce Input (95): 401 Client Error: Unauthorized for url: https://servings.my.salesforce.com/services/data/v42.0/sobjects/Task/describe/". The workflow will not run until I click into each Salesforce input and then click outside of it again to make it refresh, which takes a few seconds each time. I'm guessing this has to do with Alteryx saving an expired OAuth access token and trying to pull updated metadata.
It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive. For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox. We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility.
I’m writing about a short-coming I see in the Publish to Tableau Server Tool v2.0 (PTTS). I work in a development environment where we use different Tableau servers (i.e. development, test, production) to support product development. One of the shortcomings of PTTS is that once the Tableau server information you are connecting to is entered, validated, and the tool is configured, you can no longer ‘see’ which Tableau server/site the tool is publishing to. I think this piece of information is quite important. I know I can always us the “Disconnect” button in the tool and re-enter the information so I know which server it is pointing to, but this defeats the purpose of entering that information in the tool in the first place.
Please consider an enhancement to the tool so we developers know at a glance where (server/site) the tool is publishing. Project and Data source names are helpful, but in a development cycle, all Tableau servers may have the same Project and Data source names across all environments.
I've attached examples of the tool options when being defined and once the tool is configured – notice server URL and Site are aren’t displayed in the tool once it has been configured.