community
cancel
Showing results for 
Search instead for 
Did you mean: 

Alteryx Designer Ideas

Share your Designer product ideas - we're listening!

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the star in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

Now that we have a Snowflake Bulk Loader option, it would be great to utilize the built-in Snowflake internal staging.  This eliminates the need for an end-user to have the technical know-how or access to IT resources to utilize a separate S3 bucket and generally reduces friction in the process. 

 

There was pretty widespread support in the original Bulk Load thread: https://community.alteryx.com/t5/Alteryx-Designer-Ideas/Snowflake-Bulk-Loader/idi-p/105291/page/2#co...

As of today, for a full refresh, I can :

-create a new table

-overwrite a table. (will drop and then create the new table)

But sometimes, the workflow fails and the old table is dropped while the new one is not created. I have to modify the tool (setting "create a new table")to launch it again, which may be a complex process in companies. After that, I have to modify it again back to "overwrite".

What I want :

-create a new table-error if table already exists

-overwrite a table-error if table doesn't exist

-overwrite a table-no error if table doesn't exist (easy in sql : drop if exists...)

 

Thanks!

 

Hi Everyone,

 

Many workflows I work with along with those of my colleagues, use big databases in order to get some data. After a few steps down stream and testing, we normally just add an output and then open up that data in a new workflow to save time running the original workflow. Not that this is much of a burden, but I am used to copying and pasting tools from workflow A to workflow B, but you can't do that with the output, because in workflow B the output needs to be converted to an input. I just think it would be a cool added feature if possible. Anyone else agree?

 

Thank you,

Justin

 

 

  • Input Output

With the release of 2018.3, cache has become an adhoc task. With complex workflow and multiple inputs we need a method to cache and save the cache selection by tool. Once the workflow runs after opening, the cache would be saved at the latest tool downstream.


This way we don't have to create adhoc cache steps and run the workflow 2X before realizing the time saving features of cache.

 

This would work similar to the cache feature in 11.0 but with enhanced functionality...the best of the old cache with the new cache intent.

 

Embed the cache option into tools.

 

Thanks!

When I create a new table in a in-Db workflow, I want to specify some contraints, especially the Primary Key/Foreign Key

 

For PK/FK, the UX could be either the selection of some fields of the flow or a free field (to let the user choose a constant).

From wikipedia :

In the relational model of databases, a primary key is a specific choice of a minimal set of attributes (columns) that uniquely specify a tuple (row) in a relation (table).[a] Informally, a primary key is "which attributes identify a record", and in simple cases are simply a single attribute: a unique id. 

So, basically, PK/FK helps in two ways :
1/ Check for duplicate, check if the value inserted is legit

2/ Improve query plan, especially for join

Hello,

 

My issue is very easy to solve. I want to use the generic ODBC In database for a specific base (monetdb here but it isn't important).

 

alteryx_monetdb1_.png

 

The connexion works just fine. However, I cannot create table because the data types are changed and does not even exist. Here is my data with some Date type :

 

alteryx_monetdb2_.png

 

 

And here the error in my data stream in give me this very interesting message  :

 

Error: Entrée du flux de données (2): Erreur lors de la création de la table "formation.temp1" : [MonetDB][ODBC Driver 11.31.11]Type (datetime) unknown in: "create  table "formation"."temp1" ("AccountNumber" varchar(255),"BillToAddressID"
syntax error, unexpected IDENT in: ""Freight""

CREATE  TABLE "formation"."temp1" ("AccountNumber" varchar(255),"BillToAddressID" float,"BusinessEntityID" float,"Comment
" float,"CreditCardApprovalCode" varchar(255),"CreditCardID" float,"CurrencyRateID" float,"CustomerID" float,"DueDate" datetime,"Freight" real,"OnlineOrderFlag" float,"OrderDate" datetime,"OrderDate_Month" varchar(255),"OrderDate_Year" float,"PurchaseOrderNumber" varchar(255),"RevisionNumber" float,"SalesOrderID" float,"SalesOrderNumber" varchar(255),"ShipDate" datetime,"ShipMethodID" float,"ShipToAddressID" float,"Status" float,"SubTotal" float,"TaxAmt" float,"TotalDue" float)

 

 

1/ My field is a date, why do you want to convert it in Datetime??

2/ Datetime is not even a usual field type in sql database (at least not supported by monetdb, vertica, postgresql, oracle, etc, etc...)... it should obviously be timestamp

 

Currently, this non-specific in database ODBC connexion cannot be used at all!

 

Hi All,

 

Was very happy to see the Bulk Loader introduced for Snowflake during last release. This bulk loader is specifically available for Snowflake environments that are hosted on AWS, but does not provide functionality for those environments using Azure. As Snowflake continues to build momentum, I imagine this will be a common request. Is there something in the pipeline to add this functionality?

 

For an interim solution, we will be working toward developing some generic scripts/snowsql to mimic that bulk load, but ultimately we'd love to have this as part of the tool.

 

Best,

devKev

I really love how I can drag and drop a file directly onto the canvas from Windows Explorer and Alteryx knows to create an Input Data tool. But when I tried it with a folder today, hoping to see a Directory Input tool appear, it wouldn't do it. Could we have a similar functionality for automatically creating a Directory Input tool?

Presently when mapping an Excel file to an input tool the tool only recognizes sheets it does not recognize named tables (ranges) as possible inputs. When using PowerBI to read Excel inputs I can select either sheets or named ranges as input. Alteryx input tool should do the same.

@AdamR did a talk this year at Inspire EU about testing Alteryx Canvasses - and it seems that there is a lot we can do here to improve the product:

https://www.youtube.com/watch?v=7eN7_XQByPQ&t=1706s

 

One of the biggest and most impactful changes would be support for detailed unit testing for a canvas - this could work much like it does in Visual Studio:

 

Proposal:

In order to fully test a workflow - you need 3 things:

  • Ability to replace the inputs with test data
  • Ability to inspect any exceptions or errors thrown by the canvas
  • Ability to compare the results to expectation

To do this:

  • Create a second tab behind a canvas which is a Testing view of the canvas which allows you to define tests.   Each test contains values for one or more of the inputs; expected exceptions / errors; and expected outputs
  • Alteryx then needs to run each of these tests one by 1 - and for each test:
    • Replace the data inputs with the defined test input.   
    • Check for, and trap errors generated by Alteryx
    • Compare the output
    • Generate a test score (pass or fail against each test case)

This would allow:

  • Each workflow / canvas to carry its own test cases
  • Automated regression testing overnight for every tool and canvas

 

 

Example:

 

Testing.jpg

 

For this canvas - there are 2 inputs; and one output.

Each test case would define:

  • Test rows to push into input 1
  • Test rows to push into input 2
  • any errors we're expecting
  • The expected output of the browse tool

 

 

This would make Alteryx SUPER robust and allow people to really test every canvas in an incredibly tight way!

It would be wonderful for Alteryx to be able to connect to and query OData feeds natively, rather than using a 3rd-party driver or custom macro.   

 

OData querying is supported by quite a few familiar products, including Excel and PowerBISSIS/SSRS, FME SafeTableau, and many others. And the protocol is used to publish feeds from Microsoft Dynamics and Sharepoint, as well as many of the 10,000 publically available government datasets with API's (esp. those hosted by Socrata)   

 

I didn't see it as in the Idea section, but questions and workarounds have been discussed in the community a few times (11/15, 3/18, 4/18), and suggestions seem to be just to buy the $400-600 ODBC driver from CDATA (or ZappySys), or I could use a VBA script in Excel trigger a refresh, or create my own Alteryx connector macro (great series btw, though most was beyond my understanding!) 

   

While not opposed paying, kludging, or learning to program, they're just one more thing to build/buy, install, maintain, and break at the most inconvenient time Smiley Happy

 

Thanks,
Chadd

 

OData Overview:

OData (Open Data Protocol) is an ISO/IEC approvedOASIS standard that defines a set of best practices for building and consuming RESTful APIs. OData helps you focus on your business logic while building RESTful APIs without having to worry about the various approaches to define request and response headers, status codes, HTTP methods, URL conventions, media types, payload formats, query options, etc. OData also provides guidance for tracking changes, defining functions/actions for reusable procedures, and sending asynchronous/batch requests.  OData RESTful APIs are easy to consume. The OData metadata, a machine-readable description of the data model of the APIs, enables the creation of powerful generic client proxies and tools.

More info at at http://odata.org

Presently when mapping an Excel file to an input tool the tool only recognizes sheets it does not recognize named tables (ranges) as possible inputs. When using PowerBI to read Excel inputs I can select either sheets or named ranges as input. Alteryx input tool should do the same.

As of today, you can pass SQL from :
-input tool

-output tool

-connect in-db

The user interface is very limited, the kind of query you can pass also, welll, not very user-friendly. This generates a lot of frustration among users.

What do I suggest :
1/ A direct button "Query Builder" without having to open a new wf and drop an input box and then go with the presql tool and fight to build a query.

2/ Basically the same features than Dbeaver (https://dbeaver.io ) or DBvisualizer (https://www.dbvis.com/) or SQUIRREL http://squirrel-sql.sourceforge.net/ ):
  -Ability to pass any SQL Code I want (such as update, create, truncate, etc...) when I come from the button, "protected" sql when I am in a workflow
  -autocompletion

  -color coding (The idea is not new )

 
3/ A box "Free Sql Query" that I can branch on a indb or standard wf to pass any SQL query. The Output would be the same as input, just like it would be with a block until done.

It may be user-friendly to display the DSN description. It's usually stored in the windows registry under "description" :

 

descirption_dsn.png

 

This should display in the log and when you configure a connexion.

Directory Tool retrieves today a lot of information about a file. I must say I appreciate getting easily the size and the last write time.

But why not the owner? I have developped a macro with a powershell to do that but what a nightmare for a so little piece of information.

Hello,

 

we have several environment in our organization : dev, recept, production.

 

In order to make that change safe we intend to make several connection (standard alias)  like

alias_in_memory_pour_support.PNG

PRODUCTION_HIVE

DEV_HIVE

RECEPT_HIVE

 

In our workflows, we want to use aka:%Question.v_environment%HIVE

 

Sadly, this solution does not work despite the value defaut. 

 aka_et_alias_in_memory.PNG

 

Currently, the dynamic input tool reads variable types for two same variables in two different files as different. This causes an error while importing the files. e.g. A field called X in one file has type "double' while the same variable in other file would have type 'V_String'. The dynamic tool would give an error in this case saying that schema is different. Though its not.

 

  • Input Output

The issue : we have to regurlary change our passwords. It implies changing it for all the connections alias (at least for standard connections). 

What I'm looking for :

A commande line to execute like :
alteryxpw.exe -connection {alias name} -connection_type {user/system} -connection_mode {standard/indb} -pw {new password}

This command line could be executed by a powerhsell script.

The SQL Editor window could have a better presentation of the SQL code; two issues observed

  • First, that it's simply plain text without even a fixed-width font, much less syntax highlighting
  • Second, if you type in some manually formatted SQL code (e.g. with line feeds and indentation), then click on the "Visual Query Builder" button, then click back to the "SQL Editor" button, all the formatting is lost as it is converted to one run-on line of code, which is very difficult to read.

I understand that going between the Visual Query Builder and the SQL Editor is bound to have some issues; nonetheless the "idea" is to allow a user friendly display in the SQL Editor window:

My "implementation ideas" are based on a couple minutes with google, so hopefully this is a very feasible request; my user base is very likely to spend more time in the SQL editor than not, so this would be a valuable UX addition.  Thanks!

 

  • Input Output

I am using a Dynamic Input within a Batch Macro to allow the user to read (dump) data from multiple Oracle tables with varying schemas.  If a table read has 0 rows output, then an error message like the following is displayed at the end of the job execution:

 

capture.jpg

 

Because it is always possible to return 0 rows from the read, I'd like to mute the error message.  If this message is present, I don't want to STOP the macro/application.  I do however want to stop the application if an ERROR Message that I care about is encountered (e.g. Output file is not defined).

 

Cheers,

 

Mark

 

Top Starred Authors