This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
We have discussed on several occasions and in different forums, about the importance of having or providing Alteryx with order of execution control, conditional executions, design patterns and even orchestration.
I presented this idea some time ago, but someone asked me if it was posted, and since it was not, I’m putting it here so you can give some feedback on it.
The basic concept behind this idea is to allow us (users) to have:
Repetitive patterns to be reusable.
Select after and Input tool
Get not matching records from join
Tell Alteryx to execute some logic if something happens.
Any other condition
Order of execution
Need to tell Alteryx what to run first, what to run next, and so on…
Run this first
Execute this portion after previous finished
Wait until “X” finishes to execute “Y”
Putting all together
This approach involves some functionalities that are already within the product (like exploiting Filtering logic, loading & saving, caching, blocking among others), exposed within a Tool Container with enhanced attributes, like this example:
The approach is to extend Tool Container’s attributes.
This proposition uses actual functionalities we already have in Designer.
So, basically, the Tool Container gets ‘superpowers’, with the addition of some capabilities like: Accepting input data, saving the contents within the container (to create a design pattern, or very commonly used sequence of tools chained together), output data, run the contents of the tools included in the container, etc.), plus a configuration screen like:
Refers to the actual interface of the Tool Container.
Provides the ability to disable a Container (and all tools within) once it runs.
Idea based on actual behavior: When we enable or disable a Tool Container from an interface Tool.
Input and output data to the container’s logic, will allow to pickup and/or save files from a particular container, to be used in later containers or persist data as a partial result from the entire workflow’s logic (for example updating a dimensions table)
Based on actual behavior: Input & Output Data, Cache, Run Command Tools, and some macros like Prepare Attachment.
Order of Execution: Can be Absolute or Relative. In case of Absolute run, we take the containers in order, executing their contents. If Relative, we have the options to configure which container should run before and after, block until previous container finishes or wait until this container finishes prior to execute next container in list.
Based on actual behavior: Block until done, Cache, Find Replace, some interface Designer capabilities (for chained apps for example), macros’ basic behaviors.
Conditional Execution: In order to be able to conditionally execute other containers, conditions must be evaluated. In this case, the idea is to evaluate conditions within the data, interface tools or Error/Warnings occurrence.
Based on actual behavior: Filter tool, some Interface Tools, test Tool, Cache, Select.
Notes: Documentation text that will appear automatically inside the container, with options to place it on top or below the tools, or hide it.
This should end a brief introduction to the idea, but taking it a little further, it will allow even to have something like an Orchestration layout, where the users can drag and drop containers or patterns and orchestrate them in a solution, like we can do with the Visual Layout Tool or the Interactive Chart tool:
This has probably been mentioned before, but in case it hasn't....
Right now, if the dynamic input tool skips a file (which it often does!) it just appears as a warning and continues processing. Whilst this is still useful to continue processing, could it be built as an option in the tool to select a 'error if files are skipped'?
Right now it is either easy to miss this is happening, or in production / on server you may want this process to be stopped.
I often need to create a record ID that automatically increments but grouped by a specific field. I currently do it using the Multi-Row Formula tool doing [Field-1:ID]+1 because there is no group by option in the Record ID tool.
Also, sometimes I need to start at 0 but the Multi-Row Formula tool doesn't allow this so I have to use a Formula tool right after to subtract 1.
So adding a group by option to the Record ID tool would allow the user not to use the multi-row formula to do this and to start at any value wanted.
Love the new updates to the Browse tool in 2019.2! However, if you choose the option Open results in new window, which I do often so I can see my whole dataset, the search/filter/sort functionality goes away. Would be great if that new functionality also worked in the new window. Thanks!
When using the text mining tools, I have found that the behaviour of using a template only applies to documents with the same page number.
So in my use case I've got a PDF file with 100+ claim statements which are all laid out the same (one page per statement). When setting up the template I used one page to set the annotations, and then input this into the T anchor of the Image to Text tool. Into the D anchor of this tool is my PDF document with 100+ pages. However when examining the output I only get results for page 1.
On examining the JSON for the template I can see that there is reference to the template page number:
And playing around with a generate rows tool and formula to replace the page number with pages 1 - 100 in the JSON doesn't work. I then discovered that if I change the page number on the image input side then I get the desired results.
However an improvement to the tool, as I suspect this is a common use case for the image to text tool, is to add an option in the configuration of the image to text tool to apply the same template to all pages.
Please remove all stopwords that help to identify the sentiment of a text. E.g. words like 'no' and 'not' are currently removed when you enable the 'remove stopwords' options. Here is an example:
People will probably use the option to remove stopwords without even thinking about issues like this and might remove relevant information from texts and then do a Sentiment Analysis afterwards and wonder why the results are bad.
Dear Alteryx, please find a better stopword list or remove some words from the list.
With the new intelligence suite there is a much higher use of blob files and we would like to be able to input them as a regular input instead of having to use non- standard tools like Image, report text or a combination of directory/blob or input/download to pull in images, etc. I would like to see the standard input tool capable of bringing in blob files as well.
I have a PDF of 27 pages and each page is identical. The headers, footers and data are static in positioning on each page. It would be great if I could define the text to parse out on the first page, then that could be used to parse out all of the pages in the PDF. It would make the tool far more useful.
Instead or in addition to be able to manually enter additional stop words, it would be great if you could have an optional input connection where you could point to a file with additional stop words in it. Very manual to type the additional stop words in...
As of today, we can configure the language for all rows but it doesn't work when I have several languages in my data :
I woud like to take a field so that I can specify the language. The ideal would be a two-time configuration : globally in one language but I can overwrite that by choosing a field containing the language.
i think that for fin companies it would be very helpful to have an algorithm to analyse sentiment on various topic in articles, tweets, linktin, FB, etc. It could be helpful e.g. to understand what market think about some reg developments, projects and hot topics. Most importantly that fin companies normally deal with very spacial types of text, which are industry specific hence VADER algorithm broadly used for tweets does not really perform great on fin data. I would suggest to add FinBERT model (and BERT model as such) which are top of the pops in AI (BERT is used in goolge search engine). The pre-trained models are freely accessible. it would be very helpful if the range of model would be extended to FinBERT, for banks, FS teams, asset managers, BERT for general use, MedBERT for pharma.
I would like to share my idea that would be definitely useful for fast automation of the process with reading and correctly recognizing the text from PDF input. I wrote about that, hoping somebody has already thought about that here.
The idea is the tools "PDF Input" and "Image to Text" from "Text Mining" category to be improved, so as the text from PDF document to be read properly, no matter the text position on each page.
It could be also considered the performance of the combined tools of "PDF Input" and "Image to Text" to be improved as they work slower than the customized tool PDF Input does.
The idea also can be expanded to an entirely new tool that works out all the actions, needed for correctly reading of a PDF document without manual intervention.
In the next product version, can the parameter options for the topic modelling be changed to allow the output of both word relevance summary and interactive chart? It's a bit strange to run the tool twice to get this output.