Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
Tableau v2018.3 introduced multiple table extracts. These are particularly useful for fact table to fact table joins and fact table to entitlement table joins for row-level security where the number of rows created by the join and/or size of join results would be prohibitively large. Also they are useful for fact table to spatial joins where we might have multiple spatial objects (for example custom province/district/health facility catchment) for each row of fact table data.
So in Alteryx I'd like to be able to specify 2+ tables & their join keys and then write out a .hyper multiple tables extract.
Jonathan
At the moment if a part of your python code takes more than 30s to run, Jupyter times out and Alteryx cancels the workflow. This makes the Python Tool unusable for anything intensive and the timeout should be removed by default or be configurable per workflow.
I've made this idea as none of the solutions in these threads feel satisfactory:
Currently, you have two choices for Auto Configure while working on workflows:
Pros and Cons of both:
I would love to have something in between all, kind of an intelligent Auto Configure with following features:
This idea will add quite some complexity into the logic of Auto Configure but should have quite some potential to speed up editing workflows because network access and number of re-evaluated tools in each editing step will be reduced.
Please add a data validator workflow.
Suggested features will be the following:
1. Add validation name and set the field/s of your data you want to validate. (it can have more than one validation name on one workflow)
2. On the selected validation(name). Add features that will check/validate the information below:
A. Verify data type
B. Contains Null
C. Max and Min string length
D. Allowed values only, else it will give you an error
E. Regex expected to match and not allowed to match.
3. It can have two(2) outputs. One is True(which is match) and False(which is fail over/error).
Hi All,
Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?
For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.
Best,
devKev
When entering a number of column names in the RegEx parse mode - please can you allow either Enter or down-arrow to move down to the next cell (standard windows convention)?
Currently Enter just exists the edit mode; and down-arrow does nothing.
cc: @Hollingsworth
Where it stands now, only a file input tool can be used to pull data from Google BigQuery tables. The issue here is that the data is streamed and processed locally, meaning the power of BigQuery processing isn't actually being leveraged.
Adding BigQuery In-Database as a connection option would appeal to a wide audience. BigQuery is also standard SQL compliant with the SQL 2011 standard, so this may make for an even easier integration.
I decided to get real fancy when building a standard macro the other day. I checked the box on my macro input that made the connection optional:
It worked really well. My macro then became more complex, so I changed it to a batch macro. To my great surpise/astonishment/shock, the optional incoming connection is no longer optional:
The standard macro is working as expected on the left, but the batch macro is producing an error because my optional connection is requiring that something be connected to it.
I've been told that the code to make it optional is not there for batch macros and that this would be a product feature/improvement.
When training people on the use of action tools, something that I always have to hit on is that when you are telling the tool which piece of the XML that you are adjusting, it's sort of difficult to tell what you have selected, and super easy to accidentally select something else.
Example:
When you initially select the action to take it's this nice Blue Color. However, it still doesn't feel exactly like you have actually selected anything or told the Action Tool what to do, since it's so easy to just select any other one of these actions.
A slightly different problem is that if you are selecting an action that has been previously configured, it is just this light grey color. So it can be easy to accidentally change your settings because you may not realize it's actually set up.
Here is a recent community post that sort of outlines a few of these problems.
Please build individual *Unknown fields, one from the Left and one from the Right, into the Join tool. One *Unknown field cannot cover both side of the Join leading into the J output.
When creating a workflow I generally open a "TEMPLATE" first and then immediately save it to the "NEW WORKFLOW NAME". My template includes all my preferences that aren't set naturally within the user settings and won't get RESET by them either. It has a comment box and containers as well as logos and copyrights. It would be nice to have ready access to this feature. Maybe others have standards that they want applied to all users and their workflows too.
Thanks,
Mark
Hello all,
From https://www.sqltutorial.org/sql-triggers/
Introduction to SQL Triggers
A trigger is a piece of code executed automatically in response to a specific event occurred on a table in the database.
A trigger is always associated with a particular table. If the table is deleted, all the associated triggers are also deleted automatically.
A trigger is invoked either before or after the following event:
When you issue an INSERT, UPDATE, or DELETE statement, the relational database management system (RDBMS) fires the corresponding trigger.
In some RDMBS, a trigger is also invoked in the result of executing a statement that calls the INSERT, UPDATE, or DELETE statement. For example, MySQL has the LOAD DATA INFILE, which reads rows from a text file and inserts into a table at a very high speed, invokes the BEFORE INSERT and AFTER INSERT triggers.
On the other hand, a statement may delete rows in a table but does not invoke the associated triggers. For example, TRUNCATE TABLE statement removes all rows in the table but does not invoke the BEFORE DELETE and AFTER DELETE triggers.
So basically, I would like to create some triggers from in db tools in Alteryx.
Best regards,
Simon
Hi,
I have 2 simple ideas that would help me a little bit while working with the explorer box:
While opening certain websites, I am getting this information:
I know probably the answer to it isn't so simple, but that would give us a little bit extra flexibility while using Explorer box.
My goal is to open a word or excel file with specific documentation. If I were able to use a newer browser, I could easily open a file with a link to a webpage.
Alternatively, if I could merge a comment tool with the Explorer box tool that would also work.
It would be extremely useful to quickly find which of my many workflows feed other workflows or reports.
A quick and easy way to do this would be to export the dependencies of a list of workflows in a spreadsheet format. That way users could create their own mapping by linking outputs of one workflow, to inputs of another.
Looking at the simple example below, the Customers workflow would feed the Market workflow.
Workflow | Dependency | Type |
Customers | SQL Table 1 | Input |
Customers | SQL Table 2 | Input |
Customers | Excel File 1 | Input |
Customers | Excel File 2 | Input |
Customers | Excel File 3 | Output |
Market | Excel File 3 | Input |
Market | SQL Table 3 | Output |
It would be CRAZY AWESOME if we could get a report like this for all scheduled workflows in the scheduler.
As we do more work analyzng the canvasses that our folk are producing - it's becoming more and more necessary to have a well documented definition and schema for the XML that is used for Alteryx Canvasses.
Please could you publish the full XML definition and schema for Alteryx canvasses - this will allow groups to perform deeper analytics on how people are using Alteryx, automate quality checks; look for learning gaps; scan for dependencies etc?
Note: this relates to an idea from @dataprep here: https://community.alteryx.com/t5/Alteryx-Designer-Ideas/Documentation-tool-list-fileformat/idi-p/184...
it would be great if the formula tool could expand the intellisense to the select column box. For example, I could start typing in the select column box and it would widdle down the list of fields down.let's suppose I wanted to update field 79A, I could type in 7 and it might show something like
7
17
27
37
70
71
79A
79B.
So if I typed in 79 then, it would further reduce it to
79A
79B
And i could select 79A.
If the workflow configuration had a run for 'x' number of iterations option it would make debugging macros a lot easier. My current method consists of copying results, changing inputs and repeat until I find my problem which feels very manual.
Please update the Publish to Tableau Server connector tool to support Tableau's Ask Data feature. The data source must be recognized as an extract on Tableau Server in order for the Ask Data feature to work. Currently, all data source published using version 2.0 of the connector tool are recognized as a live data source. The work around is cumbersome and requires multiple copies of data sources to be created and managed.
User | Likes Count |
---|---|
4 | |
4 | |
3 | |
3 | |
2 |