Bring your best ideas to the AI Use Case Contest! Enter to win 40 hours of expert engineering support and bring your vision to life using the powerful combination of Alteryx + AI. Learn more now, or go straight to the submission form.
Start Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

0 Me gusta

Hi - 

 

We are using the new(ish) Anaplan connector tools; in particular, the "Anaplan Output" tool (send data TO Anaplan).

 

The issue that I'm having is that the Anaplan Output Tool only accepts a CSV file.  This means that I must run one workflow to create the CSV file, then another workflow to read the CSV file and feed the Anaplan Output Tool.

 

If it were possible to have an output anchor on the Output tool that would simply pass the CSV records through to the Anaplan Output tool, the workflows would be drastically simplified.

 

Thanks,

Mark Chappell

Pre-Filter as new option in Input Tool might reduce import data and allows to input only selected data (ie. for specific period or meeting certain conditions).

Pawel_Paleczny_1-1661942134308.png

 

Cheers,

Pawel

Add PowerPoint format file (ppt/pptx) into supported file type as direct connection in Input Tool.

Pawel_Paleczny_1-1661938082924.png

 

 

PS. I know that we have workaround allowing to import PowerPoint slides into Alteryx but I'm describing automated solution :)

 

Cheers,

Pawel

Currently the Databricks in-database connector allows for the following when writing to the database

  1. Append Existing
  2. Overwrite Table (Drop)
  3. Create New Table
  4. Create Temporary Table

This request is to add a 5th option that would execute

  • Create or Replace Table

Why is this important?

  • Create or Replace is similar to the Overwrite Table (Drop) in that it fully replaces the existing table however, the key differences are
    • Drop table completely removes the table and it's data from Databricks
      • Any users or processes connected to that table live will fail during the writing process
      • No history is maintained on the table, a key feature of the Databricks Delta Lake
    • Create or Replace does not remove the table
      • Any users or processes connected to that table live will not fail as the table is not dropped
      • History is maintained for table versions which is a key feature of Databricks Delta Lake

 

While this request was specific to testing on Azure Databricks the documentation for Azure and AWS for Databricks both recommend using "Replace" instead of "Drop" and "Create" for Delta tables in Databricks. 

 

AStasi_0-1661864644374.pngAStasi_1-1661864772827.png

Hyperion Smartview Connect

0 Me gusta

Alteryx really needs to show a results window for the InDB processes.  It is like we are creating blindly without it.  Work arounds are too much of a hassle.

We really need a block until done to process multiple calculations inDB without causing errors. I have heard that there is a Control Container potentially on the road map.  That needs to happen ASAP!!!!

Hi Team,

Microsoft Outlook is extensively used for day to day operations. With more teams, adopting Alteryx within the firm. We wanted to explore the option of end to end automation and the outputs are directly triggered to external email domains however, due to firm restrictions we cannot send email to external/internal email domains without labelling the email. Is there way to in include email labelling/classification as part of emailing tool.

Hi all,
Something really interesting I found - and never knew about, is there are actually in-DB predictive tools. You can find these by having a connect-indb tool on the canvas and dragging on one of the many predictive tools.

For instance:
boosted model dragged on empty campus:

TheOC_0-1660125228395.png


Boosted model tool deleted, connect in-db tool added to the canvas:

TheOC_1-1660125281871.png



Boosted Model dragged onto the canvas the exact same:

TheOC_2-1660125303663.png

 


This is awesome! I have no idea how these tools work, I have only just found out they are a thing. Are we able to unhide these? I actually thought I had fallen into an Alteryx Designer bug, however it appears to be much more of a feature.

 

Sadly these tools are currently not searchable for, and do not show up under the in-DB section. However, I believe these need to be more accessible and well documented for users to find.

TheOC_3-1660125345924.png



 

Cheers,
TheOC

Hello Team,

 

When we do run and cache workflow it remain in the cache memory for only the time the Alteryx designer or that workflow is open.

 

Advantage: We dont need to wait for the data to load in input data tool and the result is validated properly.

 

Disadvantage: Loss of time and money when the Alteryx Designer is closed automatically or mistaken.

 

Expectation:

 

Can we provide a token Expiry for 12 to 24 hrs so the precious time and money is not wasted.

This will be very useful when we are working on billions of Data.

0 Me gusta

By default output always print Headers as shown in item 1 below, I am looking to print item 2 as Header.  Item 3 is my output.

SyedHussaini_0-1659435340742.png

I tried item 4 below, doesn’t work.

SyedHussaini_1-1659435359699.png

Please consider this in your future releases.  It saves lot of time as the Outputs can contain hundreds of fields and output files are shared with User community who understands the Field description much more than Field name.  For example SAP Field-KUNNR does not mean nothing to a User than its description 'Customer'.

A check box on the Output Tool should able to toggle the selection between Field or its Description.

You might argue that the Rename column can be used, agree it would be difficult to manually type in hundreds of fields.  As an alternative you can provide automatic Rename population with Description.

Hello,

 

My issue is very easy to solve. I want to use the generic ODBC in memo for a specific base (monetdb here but it isn't important).

I try to ouput a flow in a MonetDB SQL database. As you can see, I only take very simple field types

simonaubert_bd_0-1659433128554.png

 

 

 

However I get this error message :

Error: Output Data (3): Error creating table "exemplecomparetable.toto": [MonetDB][ODBC Driver 11.44.0][MONETDB_SAU]Type (datetime) unknown in: "create  table "exemplecomparetable"."toto" ("ID" int,"Libellé" char(50),"Date d"
syntax error in: ""Prix""
CREATE  TABLE "exemplecomparetable"."toto" ("ID" int,"Libellé" char(50),"Date de Maj" datetime,"Prix" float,"PMP" float)


Reminder : SQL is an ISO Norm. Default type should follow it, not the MS SQL configuration. Interoperability is key

Links to : for in-db
https://community.alteryx.com/t5/Alteryx-Designer-Ideas/Generic-In-database-connection-please-stop-i...

Issues constated : MonetDB
https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Output-Data-date-is-now-datetime-makin...

Informix: https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Alteryx-date-data-type-error-trying-to...


Best regards,

Simon

Need an option in the output tool(Delete and append option is selected) so that if some error occur while writing into the table. workflow would automatically gets back to original record count. Right now if the workflow fails we are seeing zero records in the output table.

When working with .flat files, the actual Data file selection (Option #5) is a text input field.

Can we have a File Selection control here?

I think it's going to be easier than writing the whole path for a file.

Thanks!

 

 

Aguisande_1-1658673722307.png

 

 

Hi all,

 

Currently, only the Sharepoint list tool (deprecated) is working with DCM, it would be amazing to add the Sharepoint files input/output to also work with DCM.

 

 

Thank you,

Fernando Vizcaino

 

For in-DB use, please provide a Data Cleansing Tool.

 

 

01.jpg

 

0 Me gusta

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/BigQuery-Input-Error/td-p/440641

 

The BigQuery Input Tool utilizes the TableData.List JSON API method to pull data from Google Big Query. Per Google here:

 

  • You cannot use the TableDataList JSON API method to retrieve data from a view. For more information, see Tabledata: list.

 

This is not a current supported functionality of the tool. You can post this on our product ideas page to see if it can be implemented in future product. For now, I would recommend pulling the data from the original table itself in BigQuery. 

 

I need to be able query tables and views.  Not sure I know how to use tableDataList JSON API.

When loading multiple sheets from and Excel with either the Input Data tool or the Dynamic Input Tool, I usually want a field to identify which Sheet the data came from. Currently I have to import the Full Path and then remove everything except the SheetName.

 

It would be great if there was an option to output she SheetName as a field.

 

DavidP_0-1657614328613.png

 

Hello all,

I try to use Alteryx and MonetDB, a very cool column store database in Open Source.
 
When I use the Visual Query Builder, I get this :

 

simonaubert_bd_2-1657224776143.png

The fields names are totally absent.



The reason is Alteryx does not use the standard ODBC SQLColums() function at all but send some query (here "select * from demo.exemplecomparetable.fruit1 a where 0=1" ) to get a sample of data. In the same time, Monetdb sends the error "SELECT: only a schema and table name expected" (not shown to user, totally silent)

I think this should be implemented like that :
1/use of SQLColums() function which is a standard for ODBC and should work most of the time
2/if SQLColums() does not work, send the current queries.

It's widely discussed here with the MonetDB team.

https://github.com/MonetDB/MonetDB/issues/7313



Best regards,

Simon

 

 

 

When scheduled or manual Alteryx workflows try to overwrite or append the Tableau hyper extract file, it fails(File is locked by another process). This is because a Tableau workbook is using the hyper extract file. We know the behavior is expected as the file is opened by another user/process so Alteryx won't be able to modify it until the file is closed. Alteryx product team should consider this and implement it in the future to solve many problems. 

Autores con más Me gusta dados