Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
Hi everyone! I have been trying to find a way to do this without creating a new idea, but I have decided to make it an official 'Idea' to see if there is anyone else that might appreciate a feature like this (or has found there own way to do it!)
Do your workflows look like this...
but you wish they could look like this?
Well... they can with your help!
Okay, I might be crazy...but its worth a shot.
While I understand this is an extremely niche issue, in my experience, it can become very difficult to trace the data through unmanaged lines in large workflows. I think it will be great to cable manage canvas lines so workflows are easier to follow. Heck, while I am already at it, I think it we should all start calling these canvas lines cables... They don't carry electricity, but they sure do carry data!
Here is an example I created in Alteryx using select tools and containers:
Currently if I have a connection between two tools as per the example below:
I can drag and drop a new tool on the connection between these tools to add it in:
And designer updates the connections nicely, however if I select multiple tools and try and collectively drop them inbetween, on a connection then it won't allow me to do this, and will move the connection out of the way so it doesn't cause an overlap.
Therefore as a QoL improvement it would be great if there was a multi-drop option on connections between tools.
Similar to the setting that you have in many individual tools (join, append, select, et al) where you can go to options and choose to "forget missing fields" it would be nice where you could go to options for the entire flow and "forget missing fields".
This would remove the headache that you have with large flows where you make a change(s) then have to go back through each and every tool to "forget" within that tool. Yes you could still do it individually, but if you chose, you could also do it universally for the entire flow all at once to all the 'missing fields'.
In the tools that embed the "Rename" option (Select, Append Fields, Join, Join Multiple), copying the new name will copy all the information of the field configuration : tick/untick, original field name, type, size, new name and description.
Renaming the field "Rename_Field"
In my opinion, it should copy only the new name. This would be useful, especially because when you change the name of a field, it isn't automatically changed in subsequent tools, so copying it to replace it in those tools is faster than retyping it every time.
There is no tool that exists that outputs all records that are duplicates (those sharing the selected values with at least one other record) and also outputs the records that are not duplicates (those not sharing the selected values with at least one other record).
The Unique Tool is not sufficient. It only provides the first record of a unique duplicate group along with any non-duplicates and then provides a secondary output that only contains the additional records of a duplicate group. Sometimes you only care about the duplicates and want to quickly see what differs between the unique groups.
For example, if there are 4 records with the City of Austin and I am looking for duplicates on City I want to see all 4 records with Austin in the output so I can quickly compare additional fields to see what might differ, or if they are all indeed truly duplicates.
Hello all,
As of today, you can populate the Drop Down tool in the interface category with a query launched from a in-memory connection. I would really appreciate the ability to use instead an in-db connection.
Why ?
It means managing two connections instead of one, and finding ways to manage it on server for both of them, etc etc.. Simplicity is key.
Best regards,
Simon
I am working with complex workflows which use multiple files as input, located on network drives. Input tools are Input Data, Directory, Wildcard Input, Wildcard XLSX Input (from CReW macros).
Regularly, I experience very slow Designer when working on the workflows, and slow progress when running the tools mentioned above, especially when working from home. Switching off Auto Configure did not really help because I the column list sometimes does not converge even after pressing F5 multiple times, and when actively working on workflows, I have to press F5 all the time...
In order to speed up both working on workflows and running the workflows, I would like to propose a function "Cache all File Inputs" which loads and caches all file inputs at once. To achieve this state, I now have Cache and Run workflow once per every file input.
My organization use the SharePoint Files Input and SharePoint Files Output (v2.1.0) and connect with the Client ID, Client Secret, and Tenant ID. After a workflow is saved and scheduled on the server users receive the error "Failed to connect to SharePoint AADSTS700082: The refresh token has expired due to inactivity" every 90 days. My organization is not able to extend the 90 day limit or create non-expiring tokens.
If would be great if the SharePoint connectors could automatically refresh the token when it expires so users don't have to open the workflow and do it manually.
Currently there is a function in Alteryx called FindString() that finds the first occurrence of your target in a string. However, sometimes we want to find the nth occurrence of our target in a string.
FindString("Hello World", "o") returns 4 as the 0-indexed count of characters until the first "o" in the string. But what if we want to find the location of the second "o" in the text? This gets messy with nested find statements and unworkable beyond looking for the second or third instance of something.
I would like a function added such that
FindNth("Hello World", "o", 2) Would return 7 as the 0-indexed count of characters until the second instance of "o" in my string.
The idea is to have a Run option, where the workflow runs everything up to the selected tool (Like the Cache functionality does).
You select the tool, hit Run Up and the workflows executes everything "before" the selected tool.
That'll make developing much easier, specially when dealing with big workflows and constant changing data.
Not sure I'd call this a user setting, but I couldn't figure out the right heading this belongs to.
When opening files, there are often times a couple of files at that aren't run on any kind of schedule or set time frame but you come back to when you need to run them.
There should be a way to set "FAVORITES" for a handful of files that you find yourself referring to on a repeated basis, but too far back to be on the 'recents' list because you open too many other files.
When loading multiple sheets from and Excel with either the Input Data tool or the Dynamic Input Tool, I usually want a field to identify which Sheet the data came from. Currently I have to import the Full Path and then remove everything except the SheetName.
It would be great if there was an option to output she SheetName as a field.
Connecting to Smartsheets using Alteryx Desktop (and by extension, Alteryx Server) is extremely cumbersome. If a user wants to read data from Smartsheet, they are required to get an API token (preferred) or use a username/password
Then do one of the following to read data from Smartsheets:
1. a. Install a ODBC driver
b. Configure a DSN connection for ODBC
c. Use the input data using a generic ODBC connection
or
2. Use python
To write data to Smartsheets, a user can use Python or upload the data using an API call - both very hard for end users to use especially if they're not Python developers.
Regardless, all of these are problematic. On the server I manage, I have over 15 ODBC connections to Smartsheets and it's getting very hard to upgrade the server hardware because of them. Creating a native connector for input/output of data to Smartsheets will eliminate a headache of managing ODBC connections, and make it simple for Alteryx Desktop users to read and write data.
Currently it's not possible to "switch off" interface tools in the same was as we can with the other tools. This limits the functionality especially within chained apps. If we could switch these tools off it would be much easier for us to tailor the experience by allowing selections to be activated by logic, rather than simply the data therein.
Please Alteryx Gods. I beseech thee!
*lights candles*
The JOIN tool could use some love. Let's consider merging the JOIN and UNION functions into a single tool. Instead of strictly L, J, and R outputs, we could have an option to allow for all standard SQL joins:
Being able to JOIN on case-insensitive values is a big bonus (resisted urge to BOLD and change font size).
Being able to JOIN on date-range is often requested.
Being able to JOIN on numeric-range is often requested.
If we are combining tools, getting UNIQUE on L or R (or both) inputs would also save time. Most JOIN errors are because the incoming (R) data contains duplicates by KEY.
cheers,
Mark
Hi @NicoleJ
Hello all,
Like many softwares in the market, Alteryx uses third-party components developed by other teams/providers/entities. This is a good thing since it means standard features for a very low price. However, these components are very regurarly upgraded (usually several times a year) while Alteryx doesn't upgrade it... this leads to lack of features, performance issues, bugs let uncorrected or worse, safety failures.
Among these third-party components :
- CURL (behind Download tool for API) : on Alteryx 7.15 (2006) while the current release is 8.0 (2023)
- Active Query Builder (behind Visual Query Builder) : several years behind
- R : on Alteryx 4.1.3 (march 2022) while the next is 4.3 (april 2023)
- Python : on Alteryx 3.8.5 (2020) whil the current is 3.10 (april 2023)
-etc, etc....
-
of course, you can't upgrade each time but once a year seems a minimum...
Best regards,
Simon
Hello all,
I really appreciate the ability to test tools in the Laboratory category :
However, these nice tools should go out of laboratory and become supported after a few monhs/quarters. Right now, without Alteryx support, we cannot use it for production workflow.
Example given :
Visual Layout Tool introduced in 2017
https://community.alteryx.com/t5/Alteryx-Designer-Knowledge-Base/Tool-Mastery-Visual-Layout/ta-p/835...
Make columns Tool also introduced in 2017
https://community.alteryx.com/t5/Alteryx-Designer-Knowledge-Base/Make-Columns-Tool/ta-p/67108
Transpose In-DB in 10.6 introduced 2016
https://help.alteryx.com/10.6/LockInTranspose.htm
etc, etc...
Best regards,
Simon
In the Input tool, I rely heavily on the recent connection history list. As soon as a file falls off of this list, it takes me a while to recall where it's saved and navigate to the file I'm wanting to use. It would be great to have a feature that would allow users to set their favorite connections/files so that they remain at the top of the connection history list for easy access.
If the tables in the config window has lots of rows, it is quite complicated to find those of interest.
Please add a filter or search option (e.g. by the field name) to display only the relevant rows.
It would also be helpful to select or deselect multiple selected rows with one click.
Find an example from the "Select-Tool":
User | Likes Count |
---|---|
48 | |
20 | |
14 | |
13 | |
12 |