Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
I was working on the file and there are multiple sheets need to be pulled from one excel file. I was not sure how we can give one source of information and pull multiple sheet from one source as per the requirement. So wanted to submit this idea to create a toll which can pull any sheet(s) from one input tool as per requirement.
Since we can use Snowflake udfs in Alteryx, when do you think Snowflake stored procedures will be available.
When you import a csv file, I sometimes use a "TAB" as delimiter. In section 5 Delimiters I want that as an option.
I have learned that it is possible to wright "\t" but a normal choice would bed nice.
Hello all,
As of today, you can populate the Drop Down tool in the interface category with a query launched from a in-memory connection. I would really appreciate the ability to use instead an in-db connection.
Why ?
It means managing two connections instead of one, and finding ways to manage it on server for both of them, etc etc.. Simplicity is key.
Best regards,
Simon
Hello all,
As of today, you must set which database (e.g. : Snowflake, Vertica...) you connect to in your in db connection alias. This is fine but I think we should be able to also define the version, the release of the database. There are a lot of new features in database that Alteryx could use, improving User Experience, performance and security. (e.g. : in Hive 3.0, there is a catalog that could be used in Visual Query Builder instead of querying slowly each schema)
I think of a menu with the following choices :
-default (legacy) and precision of the Alteryx default version for the db
-autodetect (with a query launched every time you run the workflow when it's possible). if upper than last supported version, warning message and run with the last supported version settings.
-manual setting a release (to avoid to launch the version query every time). The choices would be every supported alteryx version.
Best regards,
Simon
Hello everybody,
I'm seeing an increased usage of geojson and kmz files across many popolar open data.
Would it be possible to add them to the supported Alteryx input formats?
Thanks!
Very beneficial will be adding extra row in Input tool (while we importing CSV) where we can define escape character. This functionality will help to avoid errors that currently occur when importing such files.
Hello all,
Big picture : on Hadoop, a table can be
-internal (it's managed by Hive or Impala, and act like any other database)
-external (it's managed by hadoop, can be shared among the different hadoop db such as hive and impala and you can't delete it by default when dropping the table
for info, about suppression on external table :
https://docs.cloudera.com/HDPDocuments/HDP3/HDP-3.1.4/using-hiveql/content/hive_drop_external_table_...
Alteryx only creates internal tables while it would be nice to have the ability to create external tables that we can query with several tools (Hive, Impala, etc).
It must be implemented
-by default for connection
-by tool if we want to override the default
Best regards,
Simon
Hello,
As of today, we can't choose exactly the file format for Hadoop when writing/creating a table. There are several file format, each wih its specificity.
Therefore I suggest the ability to choose this file format :
-by default on connection (in-db connection or in-memory alias)
-ability to choose the format for the writing tool itself.
Best regards,
Simon
Connecting to Smartsheets using Alteryx Desktop (and by extension, Alteryx Server) is extremely cumbersome. If a user wants to read data from Smartsheet, they are required to get an API token (preferred) or use a username/password
Then do one of the following to read data from Smartsheets:
1. a. Install a ODBC driver
b. Configure a DSN connection for ODBC
c. Use the input data using a generic ODBC connection
or
2. Use python
To write data to Smartsheets, a user can use Python or upload the data using an API call - both very hard for end users to use especially if they're not Python developers.
Regardless, all of these are problematic. On the server I manage, I have over 15 ODBC connections to Smartsheets and it's getting very hard to upgrade the server hardware because of them. Creating a native connector for input/output of data to Smartsheets will eliminate a headache of managing ODBC connections, and make it simple for Alteryx Desktop users to read and write data.
Right now, you can highlight 1 output tool, right click, convert to input tool. Im doing that to each of 30 outputs in 1 workflow as part of splitting a workflow that has too much spatial processing. Would be cool to highlight all of the outputs and convert all at once to save the clicker finger. Keep it rocking Alteryx!
Hello all,
We all love pretty much the in-memory multi-row formula tool. Easy to use, etc. However, the indb counterpart does not exist.
I see that as a wizard that would generate windowing functions like LEAD or LAG
https://mode.com/sql-tutorial/sql-window-functions/
Best regards,
Simon
In a similar vein to the forthcoming enhancement of being able to disable a specific output tool, my idea is to have the inverse where you can globally disable all outputs and then enable specific ones only. This should help reduce the number of clicks required/avoid workarounds using containers to obtain this functionality and allow users to be very specific in which outputs run and don't run as required.
Sometimes I need to connect to the data in my Database after doing some filtering and modeling with CTEs. To ensure that the connection runs quicker than by using the regular input tool, I would like to use the in DB tool. But is doesn't working because the in DB input tool doesn't support CTEs. CTEs are helpful for everyday life and it would be terribly tedious to replicate all my SQL logic into Alteryx additionally to what I'm already doing inside the tool.
I found a lot of people having the same issue, it would be great if we can have that feature added to the tool.
Using the Output tool to send data to a formatted spreadsheet apparently doesn't preserve formatting if the entire column is formatted. I'd love this changed to keep the formatting when its applied to an entire column. See this thread in Designer Discussions.
Creating tool which summarize all operations performed in workflow might be beneficial for people who are working on others i.e. where SQL code is required for data transformation.
While using Alteryx Desinger with below option to upload files to Azure Conatiner, storage event trigger wont kick off the adf pipeline:
There is issue with the above tool ,They are calling "CreateFile" which simply opens the file for writing, then they are uploading the file content. And finally they should be calling FlushWithClose. However, that last step is not being done.
This below microsoft thread should help Alteryx Developers to resolve the issue, many customers are looking forward to this issue and resolution.
ADF event based trigger fires when file created manually but not when databricks creates a file (mic...
Please add support for Databricks' Unity Catalog
Currently, when selecting a Databricks-connection in the “Connect In-DB”-tool, and opening the “Query Builder”, only tables in the catalog named “hive_metastore” are listed. That is, Alteryx submits the following SQL query to Databricks:
Listing tables 'catalog : hive\_metastore, schemaPattern : %, tableTypes : null, tableName : %'
However, with Unity Catalog in Databricks the namespace is three-tier and there may be multiple catalogs (and not just the "hive_metastore" catalog), see https://docs.microsoft.com/en-gb/azure/databricks/lakehouse/data-objects#--what-is-a-catalog
I reached out to Alteryx support, which replied that you currently have a feature request for implementing this change (ID TDCB-4056) and they furthermore suggested that I post here.
Thanks in advance.
When email body gets imported using latest version of the Outlook 365 tool, this tool removes the new line separators from the message body, which makes it difficult to parse relevant information out of the message body. New line separators are there prior to message being imported into Alteryx as can be verified when importing same message using different tools (for example, Python or Power Automate). Without new line separators it is not possible to accurately parse message body using Alteryx. Please add the enhancement to the Outlook 365 tool so that it doesn't remove new line separators from the message body.
This limitation of the Outlook 365 tool has been discussed in the community
I'm dealing with a database that is not always up when my production workflow runs. When the database is down, the workflow errors out because the input data connection errors. I want the workflow to gracefully execute some other steps if the database is down. I need the input data tool to allow errors to be treated as warnings, using a checkbox, and then I can have Alteryx run different logic when 0 rows are detected.
User | Likes Count |
---|---|
31 | |
7 | |
3 | |
3 | |
3 |