Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
Please add support for Databricks' Unity Catalog
Currently, when selecting a Databricks-connection in the “Connect In-DB”-tool, and opening the “Query Builder”, only tables in the catalog named “hive_metastore” are listed. That is, Alteryx submits the following SQL query to Databricks:
Listing tables 'catalog : hive\_metastore, schemaPattern : %, tableTypes : null, tableName : %'
However, with Unity Catalog in Databricks the namespace is three-tier and there may be multiple catalogs (and not just the "hive_metastore" catalog), see https://docs.microsoft.com/en-gb/azure/databricks/lakehouse/data-objects#--what-is-a-catalog
I reached out to Alteryx support, which replied that you currently have a feature request for implementing this change (ID TDCB-4056) and they furthermore suggested that I post here.
Thanks in advance.
The Output tool currently creates a zero-row file if you filter all the rows out (an empty file)
However - there's a very common need to say "don't generate a file at all if there are zero rows" - for example if you are generating a "Daily Details" file and there are no daily details today, you want to suppress this file. This frequently trips up newer users since the only way to suppress a zero-row output file today is to it with macros.
Could you add a flag to the output tool to "Suppress empty files" so that if there are no rows to be written, Alteryx does NOT create an empty file? This way the user can be in control of what to do with empty record sets?
Hello,
According to wikipedia :
A partition is a division of a logical database or its constituent elements into distinct independent parts. Database partitioning is normally done for manageability, performance or availability reasons, or for load balancing. It is popular in distributed database management systems, where each partition may be spread over multiple nodes, with users at the node performing local transactions on the partition. This increases performance for sites that have regular transactions involving certain views of data, whilst maintaining availability and security.
Well, basically, you split your table in several parts, according to a field. it's very useful in term of performance when your workflows are in delta or when all your queries are based on a date. (e.g. : my table helps me to follow my sales month by month, I partition my table by month).
So the idea is to support that in Alteryx, it will add a good value, especially in In-DB workflows.
Best regards,
Simon
Data Cleansing Tool: There should be a sub-category on the "Punctuation" cleansing. Ideally to have an option to "Include Only" or conversely "Exclude these characters" which would allow you for example to remove all characters except "." from a dollar formatted field . There are times when you need to clean almost everything except a certain punctuation or not.
I would like to suggest creating a fix to allow In-DB Connect tool's custom SQL to read Common Table Expressions. As of 2018.2, the SQL fails due to the fact that In-DB tools wrap everything in a select * statement. Since CTE's need to start with With, this causes the SQL to error out. This would be a huge help instead of having to write nested sub selects in a long, complex SQL code!
The Append Fields tool will issue a Warning if/when the Source data stream has no records that reads something like this:
Append Fields (823) There are no records present in the source.
I can imagine many situations when this issue should be flagged as a Warning. However, I have use cases when both the Source and Target data streams are expected to be empty. Because it is a common, expected scenario, I do not want it flagged as a Warning for the user.
My Idea: provide another option to suppress warnings for this situation.
Perhaps it could be a standalone checkbox, for example:
[x] Suppress Warning when both source and target streams are empty
Alternatively, the tool currently has 3 options to manage warnings or errors related to "too many" records. Perhaps this could be added as a 4th option to the dropdown list, although that would necessitate changing the label slightly.
It would be ever so helpful and save a couple extra steps if a count distinct option could be added to the crosstab tool. Seems like a slam dunk since plain ole 'count' is already a choice.
We all love seeing this. And, it's fairly easy to fix, just go find the macro and insert a new copy. But, then you have to remember the configuration and hope that it was simple.
With the tool that's there, the XML still contains the configuration, all that's missing is the tool path. It would be great to be able to right click and repair the path from the context of the missing macro.
Hi
The action of the 'tab' key in configuration window recently appears to have changed from indenting to a navigation function.
The user should be able to select which action the tab key performs.
Alternatively, tab should indent and shift-tab (or alternative) navigate. I'm not the only one who would appreciate the choice.
PuffinPanic
Hello all,
I suggest a new string function Repeat()
Repeat() forms a string consisting of the input string repeated the number of times defined by the second argument.
Repeat(text[, repeat_count])
Repeat('to',3) gives tototo
It's also a standard SQL function
https://www.w3schools.com/sql/func_mysql_repeat.asp
Best regards,
Simon
I would love to be able to see the actual curl statement that is executed as part of the download tool. Maybe something like a debug switch can be added which would produce 1 extra output field which is the curl statement itself? This would greatly enhance the ability to debug when things aren't working as expected from the download tool.
Hello there!
Currently the email tool has the following configuration:
It is a fairly easy tool to use - however one part that I would like to be improved is testing the SMTP settings - similar to how it is done on the Alteryx Server. It would be awesome to have a button as part of this page, that would send a test email, and return true/display to the user that the email sent correctly. This would stop the need to setup dummy data and a dummy output to test a connection before rolling the email tool out into a live environment/use-case.
I imagine something along these lines for this functionality:
Clicking test (and passing):
Clicking test (and failing):
Thanks,
TheOC
This year, Microsoft updated improve their API (GraphAPI) to access Office365 enviroment.
Alteryx have launched on Microsoft District in Public Gallery the Dataverse , Onedrive & Sharepoint Connectors.
Alteryx must develop as soon as possible an connector/email with same authenticator options as connectors Above and improve the emails settings.
It´s important to release and documentation to show wich permissions on Azure it's necssary to send the emails.
References:
https://docs.microsoft.com/pt-br/lifecycle/announcements/exchange-online-basic-auth-deprecated
Azure Permissions:
https://docs.microsoft.com/en-us/graph/api/user-sendmail?view=graph-rest-1.0&tabs=csharp
Parquet is a very fast, efficient and widely used data format, currently only below Parquet compression algorithms are supported and we cannot use Alteryx to read the parquet file that generated by other processes. This limits our usage in Alteryx.
Read support: Snappy and Gzip compression algorithms.
It would be great for Alteryx to support all types of Parquet format so we can maximize the use of Alteryx in data analysis.
Who needs a 1073741823 sized string anyways? No one, or close enough to no one. But, if you are creating some fancy new properties in the formula tool and just cranking along and then you see that your **bleep** data stream is 9G for nine rows of data you find yourself wondering what the hell is going on. And then, you walk your way way down the workflow for a while finding slots where the default 1073741823 value got set, changing them to non-insane sized strings, and the your data flow is more like 64kb and your workflow runs in 3 seconds instead of 30 seconds.
Please set the default value for formula tools to a non-insane value that won't be changed by default by 99.99999% of use cases. Thank you.
Hi
The wording of the tool tip displayed in results window cells with long strings is misleading. The current wording is "This cell has truncated characters".
New users tend to infer that this means that the data value has been truncated somewhere upstream. See here, here and here. Changing this message to something like "Only a portion of long strings is displayed" will help reduce the confusion immensely.
Dan
Hello all,
Change Data Capture ( https://en.wikipedia.org/wiki/Change_data_capture ) is an effective way to deal with changes in a database, allowing streaming or delta functionning. Several technos, more or less intrusive, can be applied (and combined). Ex : logs reading.
Qlik : https://www.qlik.com/us/streaming-data/data-streaming-cdc
Talend : https://www.talend.com/resources/change-data-capture/
Best regards,
Simon
for iterative macro, generally it had 2 anchors, one if it is for iterative, and it normally no output (whether got error or not)
it good to have option to remove this anchor when using it in workflow.
so other user no need to identify which one is the True output and which one is just iteration.
additional, if this can apply to input anchor.
(i just built one macro where i don't need the start input, but the input need to be iterate input)
Ability to ‘name’ the point created in the “Create Points” tool.
Instead of sticking a select tool after it to rename it from ‘centroid’ to Starting Location or Store location or whatever.
A common problem with the R tool is that it outputs "False Errors" like the following: "The R.exe exit code (4294967295) indicted an error"
I call this a false error because data passes out of the R script the same as if there were no error. As such, this error can generally be ignored. In my use case, however, my R tool is embedded within an iterative macro, and the error causes the iterator to stop running.
I was able to create a workaround by moving the R tool to a separate workflow and calling it from the CReW runner macro within my iterator, effectively suppressing the error message, but this solution is a bit clumsy, requires unnecessary read/writes, and uses nonstandard macros.
I propose the solution suggested by @mbarone (https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Boosted-Model-Error/td-p/5509) to only generate an error when the R return code is 1, indicating a true error, and to either ignore these false errors or pass them as warnings. This will allow R scripts and R-based tools to be embedded within iterative macros without breaking.
Benutzer | Anzahl |
---|---|
7 | |
5 | |
3 | |
2 | |
2 |