Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
It would be extremely helpful if Alteryx could add quick filters to the data browsing windows that would allow you to filter the contents of the window based on each column. Essentially just replicate what you can do with quick filters in Excel.
Without this, I'm often forced to copy my data out into Excel in order to do any detailed trouble-shooting, and often there's too much data to copy, which prevents me from quickly getting to what I need.
Of course you can create a whole separate filter object, but that's combersome and requires re-running the workflow.
Desperately looking for a way to connect to SQL Server Analysis Services through Alteryx as more and more of our large datasets from our older systems are moving to here in the next few months. We can connect using PowerBi with limitations (connecting 'Live' does not allow merge and connecting with Import, you need to use MDX or DAX syntax). We run into import and export limitations, too. We are not allowed access to the underlying tables, but the tables with the measures, dimensions and fields. PowerBi is a big step up from pivot tables, but Alteryx would be so much better. Ideas for connecting this up are are welcome!
Please could you enhance the Alteryx download tool to support SFTP connections with Private Key authentication as well. This is not currently supported and all of our SFTP use cases use PK.
I would like to suggest creating a fix to allow In-DB Connect tool's custom SQL to read Common Table Expressions. As of 2018.2, the SQL fails due to the fact that In-DB tools wrap everything in a select * statement. Since CTE's need to start with With, this causes the SQL to error out. This would be a huge help instead of having to write nested sub selects in a long, complex SQL code!
I understand the need for "exclusive rights" when using an input tool. Unfortunately, due to the nature of some corporate data, getting write access to a file is not always possible. I would like to have the ability to configure an input tool to open a file in "Read Only" mode while producing a warning message that the file was processed in that mode and may not contain the lastest version pf the data. I envision this as a checkbox option in the tool configuration panel.
At our organization we are required to change our passwords every few months forcing a change to my Tableau Server password. How does this relate to Alteryx? Well, every 90 days I have to change my password in the "Publish to Tableau Server Tool" for all of my workflows. This is quite a cumbersome process that could be eliminated with AD.
If you dislike manually changing your for each workflow that uses this tool then "star" this post!
Hello all,
As of today, you can populate the Drop Down tool in the interface category with a query launched from a in-memory connection. I would really appreciate the ability to use instead an in-db connection.
Why ?
It means managing two connections instead of one, and finding ways to manage it on server for both of them, etc etc.. Simplicity is key.
Best regards,
Simon
My organization use the SharePoint Files Input and SharePoint Files Output (v2.1.0) and connect with the Client ID, Client Secret, and Tenant ID. After a workflow is saved and scheduled on the server users receive the error "Failed to connect to SharePoint AADSTS700082: The refresh token has expired due to inactivity" every 90 days. My organization is not able to extend the 90 day limit or create non-expiring tokens.
If would be great if the SharePoint connectors could automatically refresh the token when it expires so users don't have to open the workflow and do it manually.
Hi,
A lot of companies now are deploying on both AWS and Microsoft Azure.
Alteryx supports AWS S3 object storage out of the box, it would be important to support Microsoft Azure blob as part of the native Alteryx product as well.
Cheers,
Adrian
Statistics are tools used by a lot of DB to improve speed of queries (Hive, Vertica, etc...). It may be interesting to have an option on the write in db or data stream in to calculate the statistics. (something like a check box for )
Example on Hive : analyse {table} comute statistics; analyse {table} compute statistics for columns;
hello,
version 2021.4 does not allow workflows to run if any of their input files are open.... would be great to have an option for the input tool that switches on/off the ability to read from open files. Some of my input files have frequent data changes and i tend to keep them open while testing/simulating results
Thank you,
abdou
Would like to direclty query Hyperion Cube / Essbase data source directly - please propose functionality in next release or add a user macro to the gallery. Thanks -cb
Alteryx has the ability to connect to data sources using fat clients and ODBC but not JDBC. If the ability to use JDBC could be added to the product it could remove the need to install fat clients.
It would be awesome if there was a cross tab in DB option because right now I have to stream out millions of records to build a cross tab.
I noticed through the ODBC driver log that Alteryx doesn't care about the kind of base I precise. It tests every single kind of base to find the good one and THEN applies the queries to get the metadata info.
Here an example. I have chosen an Hive in db connection. If I read the simba logs, i can find those lines :
Mar 01 11:37:21.318 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select USER(), APPLICATION_ID() from system.iota Mar 01 11:37:22.863 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select USER as USER_NAME from SYSIBM.SYSDUMMY1 Mar 01 11:37:23.454 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select * from rdb$relations Mar 01 11:37:23.546 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select first 1 dbinfo('version', 'full') from systables Mar 01 11:37:23.707 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select #01/01/01# as AccessDate Mar 01 11:37:23.868 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: exec sp_server_info 1 Mar 01 11:37:24.093 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select top (0) * from INFORMATION_SCHEMA.INDEXES Mar 01 11:37:24.219 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: SELECT SERVERPROPERTY('edition') Mar 01 11:37:24.423 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select DATABASE() as `database`, VERSION() as `version` Mar 01 11:37:24.635 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select * from sys.V_$VERSION at where RowNum<2 Mar 01 11:37:25.230 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select cast(version() as char(10)), (select 1 from pg_catalog.pg_class) as t Mar 01 11:37:25.415 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select NAME from sqlite_master Mar 01 11:37:25.756 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select xp_msver('CompanyName') Mar 01 11:37:26.156 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select @@version Mar 01 11:37:26.376 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: select * from dbc.dbcinfo Mar 01 11:37:26.522 INFO 5264 HardyDataEngine::Prepare: Incoming SQL: SELECT @@VERSION;
I can understand that when Alteryx doesn't know the kind of base he tries everything.. (eg : in memory visual query builder) but here, I have selected the Hive database and I have to loose more than 5 seconds for nothing.
Hello all,
According to wikipedia :
https://en.wikipedia.org/wiki/Join_(SQL)
CROSS JOIN returns the Cartesian product of rows from tables in the join. In other words, it will produce rows which combine each row from the first table with each row from the second table.[1]
Example of an explicit cross join:
SELECT *
FROM employee CROSS JOIN department;
Example of an implicit cross join:
SELECT *
FROM employee, department;
The cross join can be replaced with an inner join with an always-true condition:
SELECT *
FROM employee INNER JOIN department ON 1=1;
For us, alteryx users, it would be very similar to Append Fields but for in-db.
Best regards,
Simon
Hi,
I'm not finding it anywhere as a current option, but my company uses branded PowerPoint slides using our logo, these slides are in 16.:9 (widescreen) for slide size, but Alteryx won't output to that size even if I choose custom for page size & have Widescreen selected as an option. Could there be an Advanced Options button added that would allow users more output choices, like choosing the 16:9 ratio size output? Without it, I'm having to output the largest map I can create (13 x 9.75 in Report Map tool) and then stretch/shrink to get it to fit the 16:9 slide...for every single map/slide (currently making 40 maps at once).
Is there a work around to accomplish my goal currently? And if not, could the option be added to the Render tool? Thank you!
As per a recent discussion (https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Geopackages-Can-Alteryx-Open-GeoPackag...), please add the GeoPackage datatype to the Input tool.
For reference, the open-source project ogr2ogr has this functionality. (https://gdal.org/programs/ogr2ogr.html)
Thanks!
Connecting to Smartsheets using Alteryx Desktop (and by extension, Alteryx Server) is extremely cumbersome. If a user wants to read data from Smartsheet, they are required to get an API token (preferred) or use a username/password
Then do one of the following to read data from Smartsheets:
1. a. Install a ODBC driver
b. Configure a DSN connection for ODBC
c. Use the input data using a generic ODBC connection
or
2. Use python
To write data to Smartsheets, a user can use Python or upload the data using an API call - both very hard for end users to use especially if they're not Python developers.
Regardless, all of these are problematic. On the server I manage, I have over 15 ODBC connections to Smartsheets and it's getting very hard to upgrade the server hardware because of them. Creating a native connector for input/output of data to Smartsheets will eliminate a headache of managing ODBC connections, and make it simple for Alteryx Desktop users to read and write data.
Please add Parquet data format (https://parquet.apache.org/) as read-write option for Alteryx.
Apache Parquet is a columnar storage format available to any project in the Hadoop ecosystem, regardless of the choice of data processing framework, data model or programming language.
Thank you.
Regards,
Cristian.
User | Likes Count |
---|---|
31 | |
7 | |
3 | |
3 | |
3 |