cancel
Showing results for 
Search instead for 
Did you mean: 

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the star in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

Hello.

 

The category word  is a SQL Keyword (at least on hive). However it is put in quote ( this quote ` ) and the workflow will work without a single issue to the end. The blue color may be misleading to some users.

Bud%20Category.png

Hello,

 

Here a use case :

I work on the projects A and B with Alteryx inj IN DB mode.

 

My coworker works only on project B and have no rights to the data of project A.

 

When using temporary table in Alteryx, we both create the temporary tables in the default database. The issue is my coworker can see my temporary data of project A, which is not safe.

Solution : allow me to specify the database/schema when I create my temporary table.

These tools seem to be volatile, as in if you click on them before you run the workflow they lose their configuration. This is infuriating. Can we change this to be like every other tool where you can copy, paste or click into it at any time and it remembers its config.

 

Nick

With the amount of users that use the publish to tableau server macros to automate workflows into Tableau, I think its about time we had a native tool that publishes to Tableau instead of the rather painful exercise of figuring out which version of the macro we are using and what version of Tableau Server we are publishing to. The current process is not efficient and frustrating when the server changes on both the Tableau and Alteryx side.

Very simple. Use the wheel button on the mouse to reconfigure connections between nodes. You click on the origin or end and drop into the new anchor point.

Ok Alteryx, we totally love your product.  And I've got a super quick fix for you.  Why on earth would you Autocomplete the ubiquitous tick mark as "ReadRegistryString(Key, ValueName, DefaultValue='')"

?4-3-2018 12-08-38 PM.png

I find myself in this situation constantly where, 'dummy' suddenly becomes 'dummyReadRegistryString('HKEY_LOCAL_MACHINE\SOFTWARE\SRC\Alteryx\4.1', 'InstallDir')' the moment I strike the enter key.  

Pls help, I don't ask for much.

Hi,

 

     Currently loading large files to Postgres SQL(over 100 MB) takes an extremely long time. For example writing a 1GB file to Postgres SQL takes 27 minutes! This is serious impacting our ability to use Alteryx as an ETL tool for loading our target Postgres Data Warehouse. We would really like to see the bulk load capacity to Postgres supported by Alteryx to help alleviate the performance issues.

 

Thanks,

Vijaya

The designing interface is very slow when we design an in-db workflow.

 

 

The reason of that is that Alteryx connects everytime he needs to refresh the data. Example on Hive :

Mar 20 15:28:49.453 DEBUG 6048 HardyConnection::Connect: Default branding specific auth mech: 2
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveClientFactory::CreateClient: Create HS2 client.
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveClientFactory::GetBackendCxnPool: Create session manager.
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveClientFactory::GetBackendCxnPool: Create backend connection pool.
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveCxnPool::GetHS2Cxn: Create HS2 connection.
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveCxnPool::GetCxnFactory: Create backend connection factory.
Mar 20 15:28:49.453 DEBUG 6048 HardyHiveCxnFactory::CreateHS2Cxn: Create HS2 HTTP transport.
Mar 20 15:28:49.453 DEBUG 6048 HardySessionManager::GetSession: Getting new session handle.
Mar 20 15:28:50.399 DEBUG 6048 HardyTCLIServiceThreadSafeClient::OpenSession: TOpenSessionReq
    client_protocol = HIVE_CLI_SERVICE_PROTOCOL_V1

Maybe we could have an option on the IN DB Connection configuration to stay connected while designing (maybe with a limit time).


(PS : we also tried the option to Disable Auto Configure, it's clearly not he solution)

Current State:  When a macro contains nested macros the only method that reliably works to share them is via yxi (which I fondly refer to as my wixies). 

 

Future State:  Allow macros published to the gallery be their own tool palette so that when I or any user connects to the server the macros are there and just work, no import, no visible install just a single set of tools that work on that server. Found on Alteryx Server.jpg

Side task - also get export to yxi 

 

 

We face a big issue for our performances since we cannot as of today create tables in orc.


Connexion parameter for write :

param_ehivedb.png

Without option text file (default parameter in Simba) :

 without_option.png

without_option-2.pngwithout_option-3.png

 

With the option, the WF doesn't fail but :

 with_option-1.pngwith_option-2.pngwith_option-3.png


We want :
-to use the hdfs to write the data with data stream in
-to write the new tables with the write-indb in ORC

Alteryx Designer is an amazing data tool but it’s partner, "the Scheduler" needs some much needed upgrades. The Scheduler interface that pops up from Alteryx Designer does need a complete make over. I’m not going to address this but rather focus on the functionality that if delivered makes the Scheduler much more useful. 

 

Today I’m reading our Mongo db scheduler data using an Alteryx workflow and Tableau to show what’s happening on the Scheduler. This dashboard is what we refer to frequently to see the health of our companies data pipeline.  I’ll share both files soon. 

 

Here are the top 5 features for the Scheduler.

 

  1. Workflow priority ranking. When two or more workflows are scheduled to run at the same time, ‘priority’ sets the order of execution. Priority is set at the time of scheduling the workflow. Values range from 1 to 100. If a priority is not set the default is 50. The ‘priority’ field can be read from Mongo db. 

 

Why: At midnight we set off several workflows. We want to centrally manage which runs first based on a common ‘priority’ field. 

 

  1. Restrict which controller and workers a specific workflow can run on. Default is to run on all servers. At the time of scheduling a workflow can designated to restrict which server to execute on. This creates a field called ‘restrict’ which server it cannot execute on. 

 

Why: Some workflows only run on the main controller due to file system references. Also a worker can be tuned for CPU or Disk I/O and workflows that can benefit from this tuning.  Selecting a disk I/O intensive workflow to run on a server tuned for Disk I/O would speed up our workflows. 

 

  1. Set sequence of successfully completed workflows. 

FYI: We used the Runner tool for a short time to resolve this issue but learned quickly that the Runner tool is like a bull in a china shop and brought our server down. The runner tool as it is today is not an option for production work.

 

Why: This would allow you to run several workflows one after another. For example the first would read from a data source, the second would do calculations on the data and the third workflow would publish the data. All workflows are given a ‘workflow-number’ which can be seen in the scheduler list and read from Mongo db. 

 

  1. If a workflow fails you can set the number of attempts to run successfully. Report attempts greater than 1 in a new field called ‘attempts’ that can be read from the Mongo db. 

 

Why: Some workflows fail and if attempted to run again may work. This includes issues with locked files and workflows dependent on processes outside of Alteryx. 

 

  1. If a workflow runs more than X minutes the scheduler kills the workflow and reports a workflow error with a unique code called ‘execution-limit’.  This code can be read in the mongo db. The defaut is 90 minutes and can be set to any number of minutes. Each workflow can have it’s own limit. 

 

Why: Some workflows start to hog resources and need to be killed. If a new workflow is added this is a good way to protect the overall scheduled workflows. 

As you may know, the interrogation of Hive to get the Metadata is actually very slow on Alteryx

 

A first step of improvement (at least in the Visual Query Builder) has been proposed here

Smartest VQB

 

But the real issue for Hive is that the way Alteryx queries the Metadata : it passes "Show table" queries for all the databases. On our cluster, it means more than 400 queries that last each avout 0.5 seconds. The user has to to wait about 4 minutes.

A solution : using an API in java to ask the Hive metastore if it exists (it may be an other tab in the In database configuration). Our cluster admin has an example of a Thrift API in java that we can give you.

Result : 2 seconds for a 38700 tables in more than 500 databases !!

The idea is to store credentials, login/pw in a "credential alias".

 

Then, those credential aliases can be used in :

-traditional aliases/connection

-in database aliases/connection

-hdfs aliases/connection

-API

-on user aliases for connected controllers/gallery

...etc.

 

The idea is that I only have to change the credentials once for all the connection type (on Hive, I have the in db alias, the traditional alias and even an HDFS alias using exactly the same credentials !! and I have to change all that manually).

 

Hi,

 

A great way to promote and share the awesome Alteryx Certification process is to allow Badge and Certification URL  to show up in LinkedIn.  Here is example of IBM certification badges:

 

https://cognitiveclass.ai/badge-program/

 

Thanks

The drop down\list box have numerous ways to list values. One of them that I like is connecting to an external source. You simply have an external source file with a Name column and a Value column. It will display the data in the Name column and pass the data in the Value column. Now suppose instead of connecting to an external source I wanted to use connected tools. Currently, I would have to crosstab this data and the drop down\list box would display and pass the column names.

 

What if the drop down\list box could have an additional option added where you could connect tools and it would act identically as an external source (display the data in the Name column and pass the data in the Value column). This would be much easier and more functional!

Capture.JPG

 

We all love seeing this.  And, it's fairly easy to fix, just go find the macro and insert a new copy.  But, then you have to remember the configuration and hope that it was simple. 

With the tool that's there, the XML still contains the configuration, all that's missing is the tool path.    It would be great to be able to right click and repair the path from the context of the missing macro.

As of today, you must use a data stream out and then a hdfs tool to write a table in the hdfs in csv. Giving that the credentials are the same and that the adress in the DSN is the adress of the hdfs, it seems possible to keep the data in Hadoop and just putting it from the base to the HDFS.

Statistics are tools used by a lot of DB to improve speed of queries (Hive, Vertica, etc...). It may be interesting to have an option on the write in db or data stream in to calculate the statistics. (something like a check box for )

 

Example on Hive : analyse {table} comute statistics; analyse {table} compute statistics for columns;

 

Microsoft Office provides a facility in all its apps to make the loading of frequently used files a breeze. In the FILE OPEN function the user can "PIN" a previously opened file so that it is always easy to find and load.  This would make it easier to manage and retrieve Designer files. 

 

This is what PINNING looks like in Excel

 

pinned.png

Often we run into two pretty common situations:

  • You want to move a Select, Formula, or Filter to a different point in your workflow
  • You want to remove a tool entirely, without affecting the workflow in any other way.

In both cases, I would prefer that I can set the default behaviour of ctrl-x or delete like the way 'Delete and connect around' behaves.

 

Thanks,

Ruud

Top Starred Authors