The Sharepoint file tools are certainly a step in the right direction, but it would be great to enhance the files types that it is possible to write to sharepoint from Alteryx.
The format missing that I think is probably most in demand is pdf. If we're using the Alteryx reporting suite to create PDF reports, it would be awesome to have an easy way to output these to Sharepoint.
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Alteryx Designer is slow when using In-DB tools.
We use Alteryx 2019.1 on Hive/HortonWords with the Simba ODBC Driver configured with SSL enabled.
Here is a compare In-DB / in Memory :
We found that Alteryx open a new connection for each action :
- First link to joiner = 1 connection.
- Second ling to joiner = 1 connection.
- Click on the canevas = 1 connection.
Each connection take about 2,5 sec... It really slow down the Designer :
Please, keep alive the first connection instead of closing it and creating a new one for each action on the Designer.
Tools within a workflow needs to be able to run in parallel whereever applicable.
For example: Extracting 10 million rows from one source, 12 million rows from a different source to perform blending.
currently the order of execution is the order in which tools are dragged into the canvas. Hence Source1 first, Source2 second and then the JOIN.
Here Source1 & Source2 are completely independent, hence can be run in parallel. Thus saving the workflow execution time.
Execution time is quite crucial when you have tight data loading window.
Hopefully alteryx considers this in the next release!
Alteryx Server was recently updated to allow TLS-mediated connections to the MongoDB persistence layer. This allowed us to switch off of the embedded MongoDB to a highly-available MongoDB Atlas cluster. To our surprise after the switch, when we went to edit our workflows that make use of the persistence layer's data (Server Usage Report, etc.) to hit the new Atlas cluster, we found that the MongoDB Input tool does not support TLS connections. This absolutely needs to be changed. Based on organizational constraints, Atlas is our only option for a HA persistence layer. We absolutely have to have TLS support for the MongoDB Input tool. There is no other way for us to natively query our server persistence layer in Designer. Please bring the MongoDB Input tool into alignment with the MongoDB connections that are supported by Alteryx Server.
Often as I am scraping web sites, some clever developer has put an invisible character (ASCII or Unicode) in the data which causes terrible trouble.
I've identified 89 instances of zero-width or non-zero-width glyphs that are not visible and/or Alteryx does not classify as whitespace. There are probably more, but Unicode is big y'all.
Unfortunately, the Trim() string function only removes 4 of these characters (Tab, Newline, Carriage Feed, and Space).
REGEX_REPLACE with the \s option (which is what the Cleanse macro uses) is a little better but still only removes 20. And it removes all instances, not just leading and trailing.
I've attached a workflow which proves this issue.
@apolly: this is what I mentioned at GKO.
And I did see this post (https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Elegantly-remove-all-ASCII-characters-...), but it's too brute force. Especially as Alteryx is localized and more users need those Unicode characters.
The Multi-Field formula tool has three really powerful features that it supports:
These are really powerful within Multi-Field formulas because they allow for a dynamic process to apply across multiple fields.
However, they would also be very helpful in regular formulas and Multi-Row formulas, for code transportability.
A basic example: I have a Longitude field that is a string. I need to set it to a value of 0 if there is a null value.
My formula today:
IF ISNULL([Longitude]) THEN 0 ELSE [Longitude] ENDIF
Now lets say I want to use the same formula somewhere else, but for Latitude instead.
That formula looks like:
IF ISNULL([Latitude]) THEN 0 ELSE [Latitude] ENDIF
If I could use [_CurrentField_] instead, that would allow me to instead write both formulas as:
IF ISNULL([_CurrentField_]) THEN 0 ELSE [_CurrentField_] ENDIF
This code can easily be copied for any field that requires replacing Nulls with 0s, and doesn't require refactoring to use a Multi-Field formula instead.
This also means that if I later change my field name, the code will remain consistent. This not only speeds up development time and flexibility, but more readily allows for validation that the existing code has not changed.
With the 2019.3 release the summarize tool now includes prefixes for grouped fields. While a nice addition, in application it makes using this data downstream (like joining to other tables) more involved because of needing to remove this prefix.
It would be nice to have this as an option (a checkbox to add/remove prefixes maybe) or just revert back to pre-2019.3 behavior...thanks!
Tableau v2018.3 introduced multiple table extracts. These are particularly useful for fact table to fact table joins and fact table to entitlement table joins for row-level security where the number of rows created by the join and/or size of join results would be prohibitively large. Also they are useful for fact table to spatial joins where we might have multiple spatial objects (for example custom province/district/health facility catchment) for each row of fact table data.
So in Alteryx I'd like to be able to specify 2+ tables & their join keys and then write out a .hyper multiple tables extract.
Some of the workflows I use have multiple inputs that can take a long time to initially load. The new cache function itself has been amazing, but there is one big drawback for me: I can't cache multiple tools at the same time. Alteryx will allow me to eventually cache all of the tools I want cached, but it will take multiple times running the file. This still saves me time in the end, but it feels a bit cumbersome to set up.
With an increasing number of different projects, involving different machine learning models, it's becoming difficult to manage different package versions across workflows. Currently, the Python tool has a single virtual environment, so we need to develop models in different projects always using the same Python and package versions as the Python tool venv. While this doesn't bother the code itself too much, it becomes a problem as soon as we store and load pickled models, which are sensitive to even minor changes in packages.
This is even more so a problem when we are working on the Alteryx server, where different teams might use different packages. Currently, there is only the server admin who can install packages on the server and there can only be one version per package.
So, a more robust venv management in the Python tool would be much appreciated!
Please include IBM DB2 as an in-Database option. Currently, my primary use of Alteryx is for copying DB2 tables into Teradata for use on that server. Copying large tables and particularly joining several tables and copying the results to Teradata is too slow in Alteryx.
As a developer I want to Cache & Run all/selected inputs in a Workflow with one command. Maybe even with a keyboard shortcut?
Dear Alteryx GUI Gang,
I'll create a container and then customize the colours, margins, transparency, border and then want consistency for other containers. It would be nice to have a format painter function (brush) to apply the format of one container to another. This of course could be extended to other tools like comments. There might be a desire to apply this to more tools too, but the comments and containers would be my focus as they are almost always custom configured.
Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?
For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.
During development it would be helpful to be able to do the following in both Formula and Filter tools (and perhaps any other tool that uses custom code):
1) Highlight a line or block of code
2) Right click
Easier than manually typing or deleting "//" at every line.
Thanks in advance!
In the Gallery, the File Browse tool returns the file location on the server where the file was uploaded. This allows the file to then be read in as input to a workflow.
If you need the file path of the original file location, you have to add a Text Input for the user to manually add it.
In my case (#00293302), I used a chained app to populate a list box for the user to select the Sheet Names they would like to process through the application. Unfortunately, since I was not able to capture the original file location the application errored out. This is due to the second app using the file location on the server where the file was uploaded, which is provided by the first workflow. This file location (from the Browse tool) is a temporary file location, where inputs are immediately deleted after processing.
Want to test this out? Create an application where you Output the file path from a File Browse tool.
i know.....grrr, this doesn't match your original file location!
Can we have an option to save a workflow in a prior version for backward compatibility? I think Tableau offers this functionality.
If I have 2019.4.8 and a colleague has 2019.1.x, I cannot share my workflows because my colleague will receive a notice that the workflow was built in a newer version. I want to be able to save my workflow in 2019.1.x and send to my colleague.
This is predicated on the workflow not containing any tools/features not present in the older version. In that case, give me a warning about the specific tools/features that are not backward compatible. Thank you.
TIBCO Data Virtualization is a Data Virtualization product focused on creating a virtual data store consolidating data from throughout the enterprise. It can be accessed via a SQL query engine, and has a variety of supported connectors, including an ODBC driver.
This data source can be connected to via ODBC in Alteryx today, but error messaging is unclear/unhelpful, and attempting to use the Visual Query Builder causes Alteryx to crash.
Adding TIBCO Data Virtualization as a supported ODBC connection would empower business users to leverage this product and easily utilize this enterprise data store, enhancing the value of the Alteryx platform as a consumer of this data.
My team uses a shared macro repository (say F:\AlteryxMacros), and we recently ran into an issue with the default save location for macros. While we save most macros to our repository, there are times when folks save their macros elsewhere (let's say C:\MyAwesomeWorkflow). The issue we've encountered is that if you go to file >> save as with a macro, it will ALWAYS default to the macro repository, even when my macro is currently saved elsewhere (C:\MyAwesomeWorkflow). Speaking for a friend, people have accidentally saved things to the macro repository by accident. Or, they waste time navigating from the macro repository to the their current folder.
If a macro is saved somewhere, please change the file >> save as to default to the current folder. Thanks!
The email tool, such a great tool! And such a minefield. Both of the problems below could and maybe should be remedied on the SMTP side, but that's applying a pretty broad brush for a budding Alteryx community at a big company. Read on!
What I said the first time I ran the email tool without testing it first.
1. Can I get a thumbs up if you ever connected a datasource directly to an email tool thinking "this is how I attach my data to the email" and instead sent hundreds... or millions of emails? Oops. Alteryx, what if you put an expected limit as is done with the append tool. "Warn or Error if sending more than "n" emails." (super cool if it could detect more than "n" emails to the same address, but not holding my breath).
2. make spoofing harder, super useful but... well my company frowns on this kind of thing.
My friend @jdunkerley79 posted a terse idea: https://community.alteryx.com/t5/Alteryx-Designer-Ideas/FieldName-constant-in-Generate-Rows-Tools-an... it is inactive, but I want to extend his thoughts.
Rephrasing his idea as mine: The tool defaults the expressions to use [RowCount]. If you should either "Update Existing Field" or change the "Create New Field" the default expressions MUST be updated manually. Please update the expressions to make use of the new field.
Well, that doesn't always work! Often it will. But if you change the TYPE to date, it certainly won't. In fact, I see many questions about joining from within a DATE RANGE and the technique to build date rows from the range requires the use of DateTimeAdd(). Wouldn't it be nice (like your sample workflows) to modify the default expression based on the change of data type? I think so.
If we were thinking easy. Suppose you could have a RANGE function (dates or numerics) where you simply selected the from, to fields and gave the user the option to select the units. Now the tool auto-configures itself to create all of the "days" between the from and to dates or "1.0" and it creates all unit values between the two numeric amounts.
These would be "Alteryx" worthy enhancements in my opinion.
Most of the finance industry is using HFM or Essbase finance solutions which is flexible in keeping the data and generating reports. Alteryx should create a direct connection with Hyperion or Essbase so that it's easier to connect with the data in those systems and it can directly feed other system reports.