This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?
For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.
I didn't see it as in the Idea section, but questions and workarounds have been discussed in the community a few times (11/15, 3/18, 4/18), and suggestions seem to be just to buy the $400-600 ODBC driver from CDATA (or ZappySys), or I could use a VBA script in Excel trigger a refresh, or create my own Alteryx connector macro (great series btw, though most was beyond my understanding!)
While not opposed paying, kludging, or learning to program, they're just one more thing to build/buy, install, maintain, and break at the most inconvenient time
OData (Open Data Protocol) is an ISO/IEC approved, OASIS standard that defines a set of best practices for building and consuming RESTful APIs. OData helps you focus on your business logic while building RESTful APIs without having to worry about the various approaches to define request and response headers, status codes, HTTP methods, URL conventions, media types, payload formats, query options, etc. OData also provides guidance for tracking changes, defining functions/actions for reusable procedures, and sending asynchronous/batch requests. OData RESTful APIs are easy to consume. The OData metadata, a machine-readable description of the data model of the APIs, enables the creation of powerful generic client proxies and tools.
Right now the PublishToPowerBI connector only publishes to "My Workspace." I manage datasets that feed reports for multiple workspaces, some of which are not necessarily personal workspaces (so there is no login associated). A drop-down that lets you select which workspace, that you are a member of, would be fantastic!
The workaround right now is to ETL in Alteryx then save the dataset out to OneDrive. You can then "Publish" the Excel sheet to Power BI natively, and the data refreshes once an hour. This works for some data, but we have use cases that need refresh rates much higher than that. Plus publishing directly to Power BI would be ideal.
The Data Sources page currently lists all the different data sources that Alteryx supports - however for an administrator it's almost impossible to ensure that their designer users have the drivers for these, or are on the right version.
As an early step - can we add 1 more field to this list which points to the downloader for the driver where applicable
We are a large company with tens of thousands employees using Salesforce on a daily basis. Over the years, we have worked with Salesforce to make many customizations and create many reports to provide data for various reporting needs. However, we have increasingly found it inefficient and prone to error to download the reports manually. We have many teams using the Salesforce reports as a base to create additional business insights.
Alteryx is a great tool to manage data ETL and workflows, but it does not support pulling data from Salesforce reports directly. Instead, it only offers connectors to pull data from base Salesforce objects. The data from Salesforce objects such as tables can be useful, but do not necessarily offer the logical view of Salesforce reports, and may require a lot of efforts to reconcile the data consistency against the reports our users are used to. Sometimes, it may be impossible to repeat producing the same data from Salesforce tables as those from Salesforce reports. That in turn would cause a lot of efforts spent by the reporting teams, their audience, and users of the Salesforce reports to match things up.
Salesforce does not have any out-of-box solution to schedule downloading the reports. At our request, their support team did some research and have not found a good 3rd-party solution in the Salesforce App Exchange ecosystem that supports this need.
I strongly believe this is a great opportunity for Alteryx. Salesforce already has an API that allows for building custom applications to pull Salesforce reports. However, most Salesforce users are more business oriented and do not necessarily have the appetite to engage with their IT staff or external resources provide to develop such apps and bear the burden to main them.
I have attached the Salesforce Reports and Dashboards API Developer Guide for your reference.
It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive. For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox. We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility.
Hi, Wanted to see a connector tool for DoubleClick Search similar like Google Analytics and Adobe Connector. This would be a huge reporting benefit for a lot of digital marketing companies - worth putting effort for. Other vendors have done this in the past.
The Input Tool reads Excel however many dashboards require multiple people to update a spreadsheet. Excel is too much of a hassle to share and update. Being able to read directly from a Google Spreadsheet would allow easy multiuser access to updating spreadsheets for dashboards.
Please add Google Spreadsheets to be an option in the Input Tool.
Google Big Query can contain so-called complex queries, this means that a table "cell" can host an array. These arrays are currently not supported in the SIMBA ODBC driver for Google Big Query (GBQ), ie. you can't download the full table with arrays.
All Google Analytics Premium exports to GBQ contain these arrays, so as an example, please have a look at the dataset provided via this link:
I know there's the download but have a look at that topic, the easiest solution so far is to use an external API with import.io.
I'm coming from the excel world where you input a url in Powerquery, it scans the page, identify the tables in it, ask you which one you want to retrieve and get it for you. This takes a copy and paste and 2 clics.
Wouldn't it be gfreat to have something similar in Altery?
Now if it also supported authentication you'd be my heroes ;-)
I have used the SharePoint List Connectors with our SharePoint 2010 (on prem) instance for some time now. It works great and has become invaluable. Unfortunately, I have been told that these connectors do not work with a Cloud instance or an instance that is not on prem. We need this capability since Microsoft is pushing coporations to move to the cloud instance and there is talk that on-prem may not be available past the 2016 version that is coming soon. Many companies including mine have either completed or are close to completing a full migration which has rendered the current SharePoint Connectors useless. While this is the most important part, another piece that is missing is a SharePoint Document Library connector (similar to the Amazon S3 Download/Upload). Currently I must use the UNC path to my SharePoint folders and an easier more reliable way to save files out to OneDrive and SharePoint Online would be very beneficial.
Goal: My goals is to pull Marketo Lead Data via the Marketo Input SOAP API Connector and have specificied only 3 columns to pull for 1 day.
Issue: It appears that the Marketo Input SOAP API Connector needs to go through all 500 + columns of lead object data before it filters out the specific request I indicated. What this means is that to process 1 day of lead data for 3 columns of data, it takes 45 minutes plus and not ideal.
Client Services indicated that there is a limitation with the Marketo SOAP API tool and it would be best to utilize the Download tool and build a custom connection to the CRM System (Microsoft Dynamics) to get a predefined list of Lead IDs and then use that as a filter via the Marketo Input SOAP API connector.
Request: 1. Add Microsoft Dynamics Connector 2. Try to innovate a better Marketo Service Connector to just grab the defined fields as opposed to go through the entire processing for all records to only filter back down to.
I think it would be extremely useful to have a switch connector available in Alteryx. What I mean by a switch connector is a connecting line with an on/off state that will block the data stream through it when off. Something like below:
Switch Connector in an "Off" state
This would be extremely useful when you only want data to flow down some of the paths. In the example above, I might turn the switch connector to off because I want to see the Summarize results without outputting to a document.
The current methods for having a path/set of tools present but unused are insufficient for my needs. The two methods I and Alteryx support were able to find were:
1. Deleting the connecting line - This works, but throws up errors. Even though this is functional, it looks bad when I need to present my Alteryx module and there are errors. 2. Putting the tools in a disabled tool container - I cannot see the tools when the container is disabled. I want to be able to see my tool set-up even when I am not using it.
This is inspired by the use of switches in electrical circuit design, such as:
Please comment if you also think this would be useful, or if you have ideas for ways to improve it further. Thank you!