This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
My specific use case relates to writing to AWS but am sure there are many other use cases for federated user session token support.
Specifically, using the S3 Upload tool or Athena Bulk Write (via SIMBA and Athena ODBC), the configuration works when using a IAM user, access key, and secret access key but when using a federated user via Okta there is no option to enter the session token and authentication fails.
Alteryx desktop should support federated users' session tokens.
At our Customer we are using SAP HANA as our main database. in the moment we can only use the ODBC Client to connect against SAP HANA. In this scenario it is only possible to establish a connection with Username and Password. It would be very good if every Alteryx Designer user can connect via SSO to the HANA Database. It would be perfect to have this as standard feature and not through an add-on.
The current SharePoint API pull tool does not support the pull of managed metadata columns. It would be great if Alteryx would update the SharePoint List tools to be able to read in managed metadata columns.
With the amount of users that use the publish to tableau server macros to automate workflows into Tableau, I think its about time we had a native tool that publishes to Tableau instead of the rather painful exercise of figuring out which version of the macro we are using and what version of Tableau Server we are publishing to. The current process is not efficient and frustrating when the server changes on both the Tableau and Alteryx side.
At our organization we are required to change our passwords every few months forcing a change to my Tableau Server password. How does this relate to Alteryx? Well, every 90 days I have to change my password in the "Publish to Tableau Server Tool" for all of my workflows. This is quite a cumbersome process that could be eliminated with AD.
If you dislike manually changing your for each workflow that uses this tool then "star" this post!
It would be great if we could have a Windows Active Directory data connector tool added to the standard Alteryx toolset.
MS Excel Power Query and PowerBI both can connect to Active Directory for use as a data source, but are both very cumbersome to use. Having a connector in Alteryx that can read AD data into a workflow would be super helpful for a long list of use cases. A couple that are top of mind for me are:
-Leveraging group membership info for dynamic distribution of reports or datasets
-Being able to build reporting and dashboards about the organization (useful for Tech audit, HR, etc.)
I've seen links to an old project on GitHub of someone that started development on this, but the method (just copy these random .dlls into your program directory) is seriously frowned upon by any enterpise IT. Would be great if Alteryx could pick up that work, polish it a bit and add it to the actual Alteryx Designer toolset.
We are a large company with tens of thousands employees using Salesforce on a daily basis. Over the years, we have worked with Salesforce to make many customizations and create many reports to provide data for various reporting needs. However, we have increasingly found it inefficient and prone to error to download the reports manually. We have many teams using the Salesforce reports as a base to create additional business insights.
Alteryx is a great tool to manage data ETL and workflows, but it does not support pulling data from Salesforce reports directly. Instead, it only offers connectors to pull data from base Salesforce objects. The data from Salesforce objects such as tables can be useful, but do not necessarily offer the logical view of Salesforce reports, and may require a lot of efforts to reconcile the data consistency against the reports our users are used to. Sometimes, it may be impossible to repeat producing the same data from Salesforce tables as those from Salesforce reports. That in turn would cause a lot of efforts spent by the reporting teams, their audience, and users of the Salesforce reports to match things up.
Salesforce does not have any out-of-box solution to schedule downloading the reports. At our request, their support team did some research and have not found a good 3rd-party solution in the Salesforce App Exchange ecosystem that supports this need.
I strongly believe this is a great opportunity for Alteryx. Salesforce already has an API that allows for building custom applications to pull Salesforce reports. However, most Salesforce users are more business oriented and do not necessarily have the appetite to engage with their IT staff or external resources provide to develop such apps and bear the burden to main them.
I have attached the Salesforce Reports and Dashboards API Developer Guide for your reference.
I had a business case requiring a cost effective and quick storage solution for real time online sourced survey data from customers. A MongoDB instance would fit the need, so I quickly spun up a cluster on Mongo Atlas. Atlas was launched by MongoDB in 2016 as a database-as-a-service deployed on AWS. All instances for Atlas require TLS/SSL to connect. Currently, the Alteryx MongoDB connector does not support TLS/SSL connections and doesn't work against Atlas. So, I was left with a breakdown in my plan that would require manual intervention before ingesting data to Alteryx (not ideal).
Please consider expanding this functionality on all connectors. I am building Alteryx out in my agency as a data platform that handles sensitive customer information (name, address, email, etc.). Most tools I use to connect to secure servers today support this type of connection and should be a priority for Alteryx to resolve.
Right now the PublishToPowerBI connector only publishes to "My Workspace." I manage datasets that feed reports for multiple workspaces, some of which are not necessarily personal workspaces (so there is no login associated). A drop-down that lets you select which workspace, that you are a member of, would be fantastic!
The workaround right now is to ETL in Alteryx then save the dataset out to OneDrive. You can then "Publish" the Excel sheet to Power BI natively, and the data refreshes once an hour. This works for some data, but we have use cases that need refresh rates much higher than that. Plus publishing directly to Power BI would be ideal.
Where it stands now, only a file input tool can be used to pull data from Google BigQuery tables. The issue here is that the data is streamed and processed locally, meaning the power of BigQuery processing isn't actually being leveraged.
Adding BigQuery In-Database as a connection option would appeal to a wide audience. BigQuery is also standard SQL compliant with the SQL 2011 standard, so this may make for an even easier integration.
The join tool currently does not allow case-insensitive joins, but the find/replace tool does. Additionally- even if both sides are identical, the join tool will not join "Sean's house" to "Sean's house" because of the non-letter character in the middle. Finally - if one side is a string(2), and the other is a vString(200) - even if you have a single identical character on both sides you get uncertain outcomes unless you force the type
Please could you consider amending the join tool to include 3 new options or capabilities:
- Case insensitive join
- Allow full Unicode character set in join
- Full match across text types (irrespective of string size) - this would allow a string(2) value to match to a string(100) value as long as the string(100) value only has the same 2 characters in it as the string(2) value
That would remove a load of work from every text-join that's being done on every canvas we do.
But the real issue for Hive is that the way Alteryx queries the Metadata : it passes "Show table" queries for all the databases. On our cluster, it means more than 400 queries that last each avout 0.5 seconds. The user has to to wait about 4 minutes.
A solution : using an API in java to ask the Hive metastore if it exists (it may be an other tab in the In database configuration). Our cluster admin has an example of a Thrift API in java that we can give you.
Result : 2 seconds for a 38700 tables in more than 500 databases !!
Our company is implementing an Azure Data Lake and we have no way of connecting to it efficiently with Alteryx. We would like to push data into the Azure Data Lake store and also pull it out with the connector. Currently, there is not an out-of-the-box solution in Alteryx and it requires a lot of effort to push data to Azure.
When you right click on a Macro tool (e.g. Google Analytics) within a workflow, you can choose the version of the tool to use. However, it does not indicate which version of the tool is already in use.
Why is this an issue for me?
I have a workflow with 15 instances of the Google Analytics tool. (I needed to use the API for each month fo GA data and then use a join. I built it this way due to the restriction on the number of records.)
So when I update the Google Analytics tool I have to do it 15 times. I'd like to be sure the update is needed before I start.
Alteryx Support confirm that there is no way to tell which version of the macro is in use.
having a native PDF input tool would help me significantly. I don't have admin rights to my computer (at work) so downloading a new app to then use the "Run Command" tool is inconvenient, requires approval from IT, etc. So, it would save me (and I'm sure others) time both from an Alteryx workflow standpoint each time I need it, but also from an initial use to get the PDFtoText program installed.
It would be nice if AWS Glue had first class support in Alteryx. This would allow Alteryx to more seamlessly connect to data sources defined in the Glue metastore catalog. That alone would be handy and save on extra book-keeping. AWS Glue also has an ETL language for executing workflows on a managed Spark cluster, paying only for use. Integrating this big data tool with Alteyrx would be interesting as a way to execute in-database Spark workflows without the extra overhead of cluster management or Alteyrx connectivity