This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Who needs a 1073741823 sized string anyways? No one, or close enough to no one. But, if you are creating some fancy new properties in the formula tool and just cranking along and then you see that your **bleep** data stream is 9G for nine rows of data you find yourself wondering what the hell is going on. And then, you walk your way way down the workflow for a while finding slots where the default 1073741823 value got set, changing them to non-insane sized strings, and the your data flow is more like 64kb and your workflow runs in 3 seconds instead of 30 seconds.
Please set the default value for formula tools to a non-insane value that won't be changed by default by 99.99999% of use cases. Thank you.
Often as I am scraping web sites, some clever developer has put an invisible character (ASCII or Unicode) in the data which causes terrible trouble.
I've identified 89 instances of zero-width or non-zero-width glyphs that are not visible and/or Alteryx does not classify as whitespace. There are probably more, but Unicode is big y'all.
Unfortunately, the Trim() string function only removes 4 of these characters (Tab, Newline, Carriage Feed, and Space). REGEX_REPLACE with the \s option (which is what the Cleanse macro uses) is a little better but still only removes 20. And it removes all instances, not just leading and trailing.
Love the new updates to the Browse tool in 2019.2! However, if you choose the option Open results in new window, which I do often so I can see my whole dataset, the search/filter/sort functionality goes away. Would be great if that new functionality also worked in the new window. Thanks!
Bring back the Cache checkbox for Input tools. It's cool that we can cache individual tools in 2018.4.
The catch is that for every cache point I have to run the entire workflow. With large workflows that can take a considerable amount of time and hinders development. Because I have to run the workflow over and over just to cache all my data.
Add the cache checkbox back for input tools to make the software more user friendly.
I'd like to see Alteryx allow a second install of your license on a second, personal machine. Tableau allows this and IMO is why there is such a robust online / blog community around that product.
For those of us that work at mid-size to large organizations, there are often strict rules governing internal data and use of cloud-based data sources. If I discover some new trick I'd like the share with my fellow Alteryx analysts outside of my company, I have no clear way to do that the same way I can with Tableau where I can do it at home not using my company's data.
Being able to learn new features and test things out on commonly available public data (ever notice that Superstore data set everyone who gets Tableau has?) would accelerate what we're able to do with the community site here and the larger analytics blogging community.
You’re working on several workflows at a time and you’ve made some progress on one of them but not saved it yet. You then try open the same file again in error and somehow, as you’re trying to declarer by closing windows, you mistakenly close the workflow you’ve made all the changes to without saving, leaving the one open without all your changes.
I’d like a similar situation where for example, you get an error message if you try and open an excel workbook if it is already open on your desktop.
Right now the PublishToPowerBI connector only publishes to "My Workspace." I manage datasets that feed reports for multiple workspaces, some of which are not necessarily personal workspaces (so there is no login associated). A drop-down that lets you select which workspace, that you are a member of, would be fantastic!
The workaround right now is to ETL in Alteryx then save the dataset out to OneDrive. You can then "Publish" the Excel sheet to Power BI natively, and the data refreshes once an hour. This works for some data, but we have use cases that need refresh rates much higher than that. Plus publishing directly to Power BI would be ideal.
The admin needs to be given more control. The admin (curator) should have equal to and great control than the users of the system.
My organization is in the Healthcare industry and we have HIPAA laws to abide by when it comes to data. Not all users should be able to see all data. Developers should not have complete control over the data they publish.
Private studio -- admin should be able to control if user can publish to public gallery (compliant issues occur when users can all see and run workflows)
admin should be able to create collections so they can manage the collections and what users have access to
admin should be able to control if users can create new collections (again, compliance issues)
admin should be able to add workflows to collections (so users don't need to manage collections, admin can add the workflows for them)
admin should be able to give and remove download of workflow rights
Get tips from Tableau as they have admin controls down with their permissions process.
When I define a constant and then use it to reference an input dataset it makes it very easy for me to switch paths. It also creates a little headache for me if I copy that input or create another workflow. When I create another workflow, I must re-type and create more user constant entries. It would be nice to be able to + (create new workflow) and to generate it via a copy of the current open workflow (copy ALL preferences set plus the user constants) from the workflow. Even the ability to copy/paste the constants would be an improvement.
This got me to think a little more about localized logging options in Alteryx.
At a high level, there are ways to accomplish this in Designer at a User or System level by enabling a Logging directory and then parsing those logs with a separate Alteryx job. However, this would involve logging ALL Designer executions, which seems like it may be overkill for this need. A user can also manually save a log after each execution, although this requires manual intervention.
I think adding an option in the Runtime settings for Workflow Configuration to Enable Logging and (optionally) specify a Logging directory would be a great feature add for Designer. In my opinion this should not apply once a workflow runs on Server (Server logging should be handled in a fully standardized way), but should apply to designer "UI" execution. Having the ability to add a logging naming convention (perhaps including a workflow name and run date in the log name) would be icing on the cake.
This would allow for a piecemeal logging solution to log specific flows or processes that might be high visiblity or high importance, while avoiding saving hundreds or thousands of logs daily of less important processes, and of dev test. It would also reduce or eliminate a manual process to save these logs individually.
we use a lot the in-db tools to join our database and filter before extracting (seems logic), but to do it dynamically we have to use the dynamic input in db, which allows to input a kind of parameter for the dates, calculated locally and easily or even based on a parameter table in excel or whatever, it would be great to be able to dynamically plug a not in db tools to be able to have some parameters for filters or for the connect in-db. The thing is when yu use dynamic input in-db, you loose the code-free part and it can be harder to maintain for non sql users who are just used to do simple queries.
You could say that an analytic application could do the trick or even developp a macro to do so, but it would be complicated to do so with hundreds of tables.