Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
Problem: In certain workflows, it becomes necessary to arrange columns in a specific order for the output. While achieving the desired order for a fixed number of columns is feasible using the select tool, difficulties arise when dealing with dynamic outputs that introduce new columns during each workflow run.
Example: Consider the following scenario: the INPUT data for the select tool includes a set of Question/Answer columns. However, with every run of the workflow, new columns of this type are introduced. The challenge is to ensure that Question N and Answer N columns are grouped together in the OUTPUT dynamically. Unfortunately, this task is not easily accomplished using the current capabilities of Alteryx.
INPUT:
Company | Question 1 | Question 2 | Question 3 | Answer 1 | Answer 2 | Answer 3 |
Contoso | Blah | Bleh | Bly | N | Y | N |
DESIRED OUTPUT:
Company | Question 1 | Answer 1 | Question 2 | Answer 2 | Question 3 | Answer 3 |
Contoso | Blah | N | Bleh | Y | Bly | N |
With Python/Pandas, this problem can be easily resolved by assigning index values to each column and then sorting the columns based on the assigned index:
So, based on the Python solution, if Alteryx could do the same, it would be great. I personally think that if the Dynamic Rename tool could held the Index Info, and the select tool could also held the Sort option, this would work.
Dynamic Rename: Already can hold Description info, could hold Index Info.
Select tool: Could sort by index and hold this info when the workflow is saved.
Hope this all make sense.
Thanks.
Hi is it possible to add sheet names (to spreedsheet files) to the output of a file directory tool
Hey all,
At present, if you have an existing canvas and you want to move to a DCM Connection - you are asked something like "this will reset all of your connection details - are you sure". If you have complex queries; or pre+post SQL - then you first have to copy all of this out into Notepad before you can convert to DCM and then reconfigure it all again.
However, if you are not using DCM you can change data sources when you go into Workflow Dependancies without losing your queries etc.
Could we revisit the user experience of changing to or from a DCM connection to eliminate this "start from scratch" phenomenon - if you are converging from an existing SQL ODBC or ODB or SSVB connection to a SQL connection via DCM then it should allow you to make this conversion without losing your current configuration; and the same for any other database type.
cc: @mbarone
Please consider implementing a consistent case-sensitive option for all tools and functions.
To compare string values, including case-sensitivity: This post had a good description of the challenge, but the post has been archived:
For all the time I've used Alteryx, I thought that IF "test" = "TEST" would evaluate to false. Today I realised that isn't the case and I was surprised. I'm very surprised that "equals" performs like it does.
A few existing Ideas request case-sensitivity for individual tools:
Case insensitive option while joining two data sets
https://community.alteryx.com/t5/Alteryx-Designer-Desktop-Ideas/Case-insensitive-option-while-joinin...
Unique tool enhancement - deal with case sensitive data
https://community.alteryx.com/t5/Alteryx-Designer-Desktop-Ideas/Unique-tool-enhancement-deal-with-ca...
This new Idea requests system-wide consideration for case-sensitivity, for all tools and functions.
Current state:
These tools and functions are case-sensitive:
These tools and functions are NOT case-sensitive:
These tools and functions can be either case-sensitive or NOT case-sensitive, depending on the options used:
Current Challenges:
How do we easily identify Lower Case, Upper Case, Mixed Case?
How do we easily compare strings for equality, using case sensitivity?
Request:
Ensure all tools and functions include an option to ignore or consider Case
Create new functions for IsUpperCase, IsLowerCase, IsMixedCase
Create a new function for IsEqual, with an option to ignore or consider Case
See attached workflow, which
Most of the time I don't want/need the column that I parsed. Provide a check box for if you want the root column output.
Hi!
Just thought up a simple improvement to the US Geocoder macro that could potentially speed up the results. I'm doing an analysis on some technician data where they visit the same locations over & over again. I'm doing a full year analysis (200k + records) & the geocoder takes a bit to churn thru that much data. In the case of my data though, it's the same addresses over & over again & the geocoder will go thru each one individually.
What I did in my process & could be added to the macro is to put a unique tool into the process based off address, city, state, zip, then Geocode the reduced list, then simply join back to the original data stream using a join based off the address, city, state, zip fields (or use record id tool to created a unique process id to join off).
In my case, the 200k records were reduced to 25k, which Alteryx completed in under a minute, then joined back so my output was still the 200k records (all geocoded now).
Not everyone will have this many duplicates, but I'd bet most data has a few, & every little bit of time savings helps when management is waiting on the results haha!
Please update the Render tool to allow users to name the Excel sheet for the output. Alteryx currently errors when using same naming convention that works in normal Output tool.
When a user wants to use the find nearest to say find the nearest within 200 miles the dropdown stops at 100.
Similar if they want a number in between IE 15 the interface is not intuitive.
While you can just type the number in the interface doesn't look like you are able to.
Simply adding a "Custom" selection at the bottom would make this much more intuitive.
Hi,
Would be helpful to have an Input and Output Tool for ProjectOnline like the SharePoint and OneDrive Tools.
This way we can read the projects in a tabular form and automate our project management tasks.
Thank you.
When you start using DCM - you may have existing canvasses which use regular old connection strings which you want to migrate to DCM.
Currently (in 2023.1.1.123) - when you select "Use Data Connection Manager" - it shreds the configuration of your input tool which makes it difficult to just convert these from an existing connection to a DCM connection
The only way to then make sure that you don't lose any configuration on the tool then is to use the XML editing functionality of the tools and copy across your old configuration.
Could you please add the capability to keep my current tool configuration, but just change from using a regular old connection string to using DCM?
Many thanks
Sean
cc: @wesley-siu @_PavelP
I have developed many workflows, macros, and apps, and I have always had to find a workaround for displaying information on the user config page or user interface.
For example, I want to input 'Default text' into the Text Box interface tool, but the problem is that it does not accept any external connection.
It would be great if this tool had a Q input anchor that could accept data from a connected tool (in both single or multi-line mode) or from external input (such as a file for DropDown list or List Box tools).
As an international organization we deal with clients in multiple-countries.
Name matches for names including Chinese characters generate a unicode conversation warning and are excluded from the fuzzy match.
It would be good if fuzzy match could be enhanced to handle Chinese characters.
After using the PCA can there be a model object to output to be able to "score" new data?
Similar to PCA transform here https://stackoverflow.com/questions/26182329/how-do-i-convert-new-data-into-the-pca-components-of-my...
As currently there is no way to use this model with new data
It would be great if you could include a new Parse tool to process Data Sets description (Meta data) formatted using the DCAT (W3C) standard in the next version of Alteryx.
DCAT is a standard for the description of data sets. It provides a comprehensive set of metadata that can be used to describe the content, structure, and lineage of a data set.
We believe that supporting DCAT in Alteryx would be a valuable addition to the product. It would allow us to:
We understand that implementing support for this standards requires some development effort (eventually done in stages, building from a minimal viable support to a full-blown support). However, we believe that the benefits to the Alteryx Community worldwide and Alteryx as a top-quality data preparation tool outweigh the cost.
I also expect the effort to be manageable (perhaps a macro will do as a start) when you see the standard RDF syntax being used, which is similar to JSON.
DCAT, which stands for Data Catalog Vocabulary, is a W3C Recommendation for describing data catalogs in RDF. It provides a set of classes and properties for describing datasets, their distributions, and their relationships to other datasets and data catalogs. This allows data catalogs to be discovered and searched more easily, and it also makes it possible to integrate data catalogs with other Semantic Web applications.
DCAT is designed to be flexible and extensible, so they can be used to describe a wide variety. They are both also designed to be interoperable, so they can be used together to create rich and interconnected descriptions of data and knowledge.
Here are some of the benefits of using DCAT:
Here are some examples of how DCAT is being used:
As the Semantic Web continues to grow, DCAT is likely to become even more widely used.
DCAT
RDF
When creating a connection using DCM (example being ODBC for SQL) - the process requires an ODBC Data Source Name (see screenshot 1 below).
However, when you use the alias manager (another way to make database connections) - this does allow for DSN-free connections which are essential for large enterprises (see screenshot 2 below).
NOTE: the connection manager screens do have another option - Quick Connect - which seems to allow for DSN-free connections, but this is non-intuitive; and you're asked to type in the name of the driver yourself which seems to be an obvious failure point (especially since the list of all installed drivers can be read straight from the registry)
Please could we change DCM to use the same interfaces / concepts as the alias screens so that all DCM connections can easily be created without requiring an ODBC DSN; and so that DSN-free connections are the default mode of operation?
Screenshot 1: DCM connection:
screenshot 2
cc: @wesley-siu @_PavelP @ToddTarney
I would love a tool to be created for looking up a value in a table based on a condition. It could be called "Lookup." One input to the tool would be the lookup list, the other is the main database. Inside the tool you could enter functions that can query the lookup table and return the results either as an overwrite of an existing field in the main DB or as a new field in the main DB, similar to the options in the Multi-Row Formula tool.
Here is a link to my post in Community that explains the problem. The solution, in a nutshell, was to create a Join (which resulted in millions of additional rows), run the conditional formula, then filter to get rid of the millions of rows that were created by the Join so only those that met the condition remained (the original database rows).
Here is the text of my Community post describing my project (slightly modified for clarity):
Table 1: A list of Pay Dates (the lookup table)
Table 2: Daily timekeeper data with Week Start and Week End Date fields.
The goal: To find the Pay Date in Table 1 that is greater than the Week Start Date in Table 2 and no more than 13 days after the Week End Date in Table 2.
[Table 2: Week Start Date] < [Table 1: Pay Date]
and [Table 2: Week End Date] < [Table 1: Pay Date]
and DateTimeDiff([Table 1: Pay Date], [Table 2: Week End Date], 'Days') <= 13
There are many different flows I could use this type of tool for that would save time and simplify the flow.
Thanks!
Sometimes, Control Containers produce error messages even if they are deactivated by feeding an empty table into their input connection.
(Note that this is a made up example of something which can happen if input tables might be from different sources and have different columns so that they need separated treatment.)
According to the product team, this is expected behaviour since a selection does not allow zero columns selected. This might be true (which I doubt a bit), but it is at least counter-intuitive. If this behaviour cannot be avoided in total, I have a proposal which would improve the user experience without changing the entire workflow validation logic.
(The support engineer understands the point and has raised a defect.)
Instead of writing messages inside Control Containers directly to the log output (on screen, in logfile) and to mark the workflow as erroneous, I propose to introduce a message (message, warning, error) stack for tools inside Control Containers:
This would result in a different sequence of messages than today (because everything inside activated Control Containers would be reported later than today). Since there’s no logical order of messages anyways, this would not matter. And it would avoid the apparently illogical case that deactivated Control Containers produce errors.
Our company has a need to link a new data source in Athena. We have been able to establish a connection using the input functionality however the connection is so slow it is unusable. We need to have Alteryx build an In Database option for Athena to allow us to link our data lake to Alteryx.
This is a hybrid idea related to both posts regarding dynamic tool configuration during runtime / without having to run an analytic app.
What I would like to propose is a new optional connection type for the interface tools that can be updated with incoming connections (having a Q letter with white background), namely Drop Down, List Box, Tree and Map tools. This could be a simple R letter in a square for example, which would be located to the left of the incoming question anchor.
Use Case
Imagine an app where there are two control containers and three interface tools (Action tools excluded from the count) outside those containers, one of them is a Text Box connected to a filter tool (via an Action tool) in the first control container with the purpose of limiting the dataset by specifying a city for example, another one is a Numeric Up Down for limiting the dataset by the average transaction amounts that are greater than the specified amount. These two interface tools are contained in a Group Box in the Interface Designer.
The third interface tool is a Drop Down tool which obtains the values (which will be Store Name for this example) from the results of the Select tool (in the second control container that is connected to the output anchor of the first control container) that is connected to an incoming filter tool which is modified by the previously mentioned interface tools. Output anchor of this Select tool is connected to the hypothetical R anchor on the top of the Drop Down tool, which is then connected to an outgoing filter tool that is connected to a series of tools which ends with a Browse tool that displays basic KPI information for the store specified from the Drop Down tool.
The main difference of the R (Refresh) anchor from the Q anchor is that it will enable the user to dynamically update the incoming values (i.e., choices for a drop down tool) without having to run the workflow. Alteryx Designer will automatically execute only the tools necessary to be able to update the values (up to a certain point of the workflow only, which may also be indicated by the boundaries of the control containers containing the target tool) for the R anchor connected applicable Interface tools specified above. This will be possible by clicking the hypothetical confirm button (same appearance with the Apply Data Manipulations button) which only appears next to the Interface tools (or the Group Boxes containing them instead) that are automatically determined by Alteryx Designer to be providing downstream data to the the tools (T anchor of the Filter tool for example) sending values to the applicable Interface tools having an incoming R anchor connection.
I saw that a similar feature recently became available with Alteryx Analytics Cloud Platform with the App Builder product, and I think that Alteryx Designer Desktop could definitely benefit both from this feature and additional App Builder features (that can be adapted to Desktop counterpart) in the upcoming releases.
Apologies if this has been suggested or exists. I often find myself using manual Excel files as a data source. These files frequently use cell formatting elements, such as cell color and text color, to convey important information. However, when these files are imported into Alteryx, this valuable formatting information is unfortunately lost.
To address this, a dedicated input tool that can read Excel files with separate fields for these formatting elements would be very helpful. This would be incredibly beneficial, especially when the data lacks other fields that relate to the coloring. Currently, I manage to achieve this using a Python tool, but integrating this as a built-in feature in Alteryx would undoubtedly be more efficient and user-friendly. This enhancement would not only simplify data preparation but also ensure the preservation of the full context of the original Excel file.
User | Likes Count |
---|---|
7 | |
7 | |
5 | |
3 | |
3 |