This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Many of the errors displayed in the 'results/messages' section in designer when a workflow fails are vague and unhelpful. For example, I had a workflow recently fail because my data source table had duplicate rows. The error reported was "the pipe has ended". This is a frequent problem when debugging workflow issues, and a lot of extra time is needed due to lack of useful error messages. Please enhance this error logging function to attach to underlying data source systems, etc. so that better error messages can be displayed when a workflow fails.
Note: Depending on database platform/function/application where an error occurs, some errors are better than others. However, we work with the Snowflake database platform frequently, and many of the errors we've seen while testing workflows are unusable. If different teams support the error logging functionality depending on source system, etc. where the error occurred, we would request the Snowflake logging to be enhanced first if possible.
I have been developing and accumulating custom functions over the years and they have proved to be very useful. I am submitting these here. I hope they are found to be beneficial.
Functions included in the attached file include:
StandardDate(String) - Transforms any valid string to the standard date format yyyy-mm-dd
FileDirDepth(Path) - Returns the zero based depth of the path (zero being the root)
FileGetFolder(Path, Depth) - Returns the folder name given the zero based depth in the path (zero being the root)
LeftPart(String, Separator) - Returns the left part of a string up to the first separator
RightPart(String, Separator) - Returns the right part of a string after the first separator
Split(String, Delimiter, Index) - Returns the zero indexed part of a delimited string
CleanSpace(String) - Trims string and replaces multiple spaces with a single space
UnicodeToASCIIBasic(String) - Replaces all Unicode Characters with ASCII Basic equivalents
InList(Variable, List) - If Variable is in List returns True. List must be pipe delimited
IsValidEmail(String) - Returns True if string is a valid email format
IsUUID(String) - Returns True if string is a valid UUID
To make these functions available in Alteryx, place the attached xml file in the folder C:\Program Files\Alteryx\bin\RuntimeData\FormulaAddIn if you have a standard installation. If the install is non-standard, find the \bin\RuntimeData\FormulaAddIn folder and place the attached xml file there. Ateryx will need a restart for the functions to be available.
It is just a bit of annoyance, really. I'd like to see the option of inputting a hexcode of color and/or a screen color picker in the color dialog. At the moment, you have to change R, G, B separately or play around with the cursor to find the right color.
The color dialog is relevant for the documentation purposes but also reporting tools and I'm sure it would make life easier to some people, especially when branding colours are important.
Who needs a 1073741823 sized string anyways? No one, or close enough to no one. But, if you are creating some fancy new properties in the formula tool and just cranking along and then you see that your **bleep** data stream is 9G for nine rows of data you find yourself wondering what the hell is going on. And then, you walk your way way down the workflow for a while finding slots where the default 1073741823 value got set, changing them to non-insane sized strings, and the your data flow is more like 64kb and your workflow runs in 3 seconds instead of 30 seconds.
Please set the default value for formula tools to a non-insane value that won't be changed by default by 99.99999% of use cases. Thank you.
Well, the title is pretty simple : it appears that the tendancy right now is to have web version of any software on a server.
A few notes about that : -a lot of Alteryx competitors are already in this mode and it's hard to sell you're still with a desktop-only mode for design, even if the product is far better. -a good idea is the one used by Qlik with Qlik Sense : they still have a desktop and a web version of Sense but the desktop works mainly as an hidden browser plus an engine. The web version is cool too because you can make your own application, or your own data connection etc.. -the main interest of a web implementation of Alteryx would be to reduce installation on client computers (and that means packaging the installer, managing the data connection, the paths, the access to macros... etc) and to have a better control of the users.
PS : this idea is soooo simple and so obvious I'm surprised I didn't find it. It may be a duplicate.
Love the new updates to the Browse tool in 2019.2! However, if you choose the option Open results in new window, which I do often so I can see my whole dataset, the search/filter/sort functionality goes away. Would be great if that new functionality also worked in the new window. Thanks!
Sometimes formulas get pretty long. There are cases of deeply nested conditionals, concatenation of long strings, cases where multiple casts and parses are used, etc. where formulas get pretty large and unwieldy. The current system of wrapping lines and managing the size of the properties pane can be a hassle, especially if you are trying to use any sort of whitespace formatting to make the formulas more readable.
My solution is this is pretty simple, add a pop-out window for formulas. It could be a context menu option from right-clicking the formula box itself, a button on the bar at the top of each formula, or any number of other things.
A really good example of this is MS Access. You can right-click any text box that takes an expression and open it in the expression editor pop-up window. The current system is more like excel where you're stuck with whatever box size you're given.
Currently, when one uses the Google BigQuery Output tool, the only options are to create a table, or append data to an existing table. It would be more useful if there was a process to replace all data in the table rather than appending. Having the option to overwrite an existing table in Google BigQuery would be optimal.
You’re working on several workflows at a time and you’ve made some progress on one of them but not saved it yet. You then try open the same file again in error and somehow, as you’re trying to declarer by closing windows, you mistakenly close the workflow you’ve made all the changes to without saving, leaving the one open without all your changes.
I’d like a similar situation where for example, you get an error message if you try and open an excel workbook if it is already open on your desktop.
I'd like to see Alteryx allow a second install of your license on a second, personal machine. Tableau allows this and IMO is why there is such a robust online / blog community around that product.
For those of us that work at mid-size to large organizations, there are often strict rules governing internal data and use of cloud-based data sources. If I discover some new trick I'd like the share with my fellow Alteryx analysts outside of my company, I have no clear way to do that the same way I can with Tableau where I can do it at home not using my company's data.
Being able to learn new features and test things out on commonly available public data (ever notice that Superstore data set everyone who gets Tableau has?) would accelerate what we're able to do with the community site here and the larger analytics blogging community.
With the release of 2018.3, cache has become an adhoc task. With complex workflow and multiple inputs we need a method to cache and save the cache selection by tool. Once the workflow runs after opening, the cache would be saved at the latest tool downstream.
This way we don't have to create adhoc cache steps and run the workflow 2X before realizing the time saving features of cache.
This would work similar to the cache feature in 11.0 but with enhanced functionality...the best of the old cache with the new cache intent.
A minor, but time saving GUI enhancement would be appreciated. When adding a tool to the canvas, the current behavior is to make visible the tool anchor that was last used on prior tools. That being said, when I look at the results window, I might be adding a "vanilla" configuration tool to the canvas and stare at a BLANK results window. When users are adding tools to the canvas, I suggest that the best practice is to VIEW the incoming data before configuring the tool.
I ALWAYS set the results to view the INCOMING DATA ANCHOR.
Currently the InDB tools require to select a DSN that is defined on your computer.
This makes any workflow which uses a DSN incredibly painful to deploy to the server, since the DSN needs to be created on every worker node and for a large server environment this can mean creating DSN entries on 6+ worker nodes in prod plus prod server plus dev/UAT environment plus dev/UAT worker nodes.
Could we please change the InDB tools to default to DSN-less connections, where the connection persists the connection details in-line so that it can deploy to the server without a DSN setup (since all connection details are contained within the connection string)?
When saving a workflow to the gallery, the manage assets box is a fixed size. Due to the length of our file paths, the full path name is cut off. This makes it difficult when trying to package macros to select the correct box. I understand you can group assets by tool, however with the more complicated workflows this can still present it's challenges. I'd like to see this box to be re-sizeable.
In user settings you can define a "Logging Directory" and if you do the system will send the Output Log (Results view messages) to a file in that folder. The name generated is Alteryx_Log_ + an apparent sequential number, example: Alteryx_Log_1519833221_1.
This makes it impossible to identify which flow it is associated with and which instance of execution simply by looking at the name, you have to parse the content to see the flow name and start/end timestamps. For trouble shooting we want to be able to look at the list of file names and quickly see which file, of possibly hundreds of files, we need to look at to see what went wrong.
We build some pretty robust maps with multiple connections and it would be great to copy the map tool and paste it with all of the connections when we want to tweak the map slightly but keep our original map. It is a regular occurrence for us to have a very detailed map grouping by trade area name and then may want to have an overview map with all of the same connections but slightly different layout. Tracking down the connections, reconnecting them and naming them accordingly takes a substantial amount of time even in the most organized of workflows. This function would be a huge time-saver. It would also be of value with joins and unions - anywhere you have multiple streams coming in.
All the items in the 'View' menu have useful shortcuts. For instance, I'm always using Ctrl+Alt+D for the interface designer.
It would be nice to be able to quickly check workflow dependencies; at the moment there are far too many clicks to get there given the amount they are accessed (by myself and my colleagues at least...)
Quite often the window is only required briefly to check whether absolute/relative paths are used and to see inputs/outputs. It would really improve speed and ease of use if the window could be brought up with a couple of key strokes.