When you use Create Points tool - you then almost always need to use a Select tool to rename that point.
Can we please add a single text field to the Create Points tool - which would then allow us to create and name a point in one step?
This has probably been mentioned before, but in case it hasn't....
Right now, if the dynamic input tool skips a file (which it often does!) it just appears as a warning and continues processing. Whilst this is still useful to continue processing, could it be built as an option in the tool to select a 'error if files are skipped'?
Right now it is either easy to miss this is happening, or in production / on server you may want this process to be stopped.
I surprisingly couldn't find this anywhere else as I know it's been discussed in person on many occasions.
Basically the Formula tool needs to be smarter in many ways, but this particular post focuses on the Data Type component.
The formula tool, should not always default to V_String as the data type when entering data or a formula into the formula tool, it should look at the data type and estimate the most likely option.
I know there are times where the logical type might not be consistent in all fields, but the Data Preview and the Function of the formula should be used to determine the most likely option.
E.G. If I type a number or a date directly into the formula tool, then Alteryx should be smart enough to change the data type from the standard V_String to Int, Double or date.
This is an extension to the ideas posted here:
Improve Help Documentation or in-tool options for handling null values in statistical tools like Weighted Average or Linear Regression. For instance, checkbox to remove null value records, or at least warn users.
In the processing of learning to perform linear regression in RStudio and Alteryx, I came across differing outputs depending on how null values were addressed. Take the Weighted Average tool for example.
In R, the weighted.mean function treats null values in the variable of interest as if they were not there. If the user does not specify that null values exist, the result is NA. If any null values exist in the weight field, the result is NA.
Since I am more familiar with Alteryx, I originally did the data preparation—including calculating the weighted means—in Alteryx. When comparing these weighted means with those generated in R, I found that Alteryx treats the null values as zeros (i.e. includes them in the calculation). The user would have to know this is incorrect and first filter out the null values. See screenshot examples.
This is also the case within the Linear Regression tool. If null values are not omitted prior to regression, the results are wildly different. Perhaps this is known by more experienced users/statisticians, but this incorrect usage would have gone on unbeknownst to be had I not cross-checked with RStudio.
Recently I was helping a client to design the workflow to do transformation. In the middle of the work, I feel a bit lost on handling so many fields and thinking it would be great if there is a feature that allow me to track the field actions along the workflow. It could be something like a configuration on the canvas, user activate it only when they want to.
And when it is activated, the workflow could become:
So it is easier to find the path of certain field along the whole workflow.
Or is there any method to achieve this at the moment?
It would be a given wherein IT would have invested effort and time building workflows and other components using some of the tools which became deprecated with the latest versions.
It is good to have the deprecated versions still available to make the code backward compatible, but at the same time there should be some option where in a deprecated tool can be promoted to the new tool available without impacting the code.
Following are the benefits of this approach -
1) IT team can leverage the benefits of the new tool over existing and deprecated tools. For e.g. in my case I am using Salesforce connectors extensibly, I believe in contrast to the existing ones the new ones are using Bulk API and hence are relatively much faster.
2) It will save IT from reconfiguring/recoding the existing code and would save them considerable time.
3) As the tool keeps forward moving in its journey, it might help and make more sense to actually remove some of the deprecated tool versions (i.e. I believe it would not be the plan to have say 5 working set of Salesforce Input connectors - including deprecated ones). With this approach in place I think IT would be comfortable with removal of deprecated connectors, as they would have the promote option without impacting exsiting code - so it would ideally take minimal change time.
In addition, if it is felt that with new tools some configurations has changed (should ideally be minor), those can be published and as part of
promotions IT can be given the option to configure it.
In SQL, obtaining partial sums in a grouped aggregation is as simple as adding "WITH ROLLUP" to the GROUP BY clause.
Could we get a "WITH ROLLUP" checkbox in the Summary Tool's confg panel in order to produce partial sums?
(This was initiated as a question here.)
Here is an example of without rollup vs. with rollup, in SSMS:
(Replacing "NULL" with the phrase "(any)" or some such, and we have a very useful set of partial sums.
It would help if there is some option provided wherein one can test the outcome of a formula during build itself rather than creating dummy workflows with dummy data to test same.
For instance, there can be a dynamic window, which generates input fields based on those selected as part of actual 'Formula', one can provide test values over there and click some 'Test' kind of button to check the output within the tool itself.
This would also be very handy when writing big/complex formulas involving regular expression, so that a user can test her formula without having to
switch screens to third party on the fly testing tools, or running of entire original workflow, or creating test workflows.
When CrossTab is used, string data in fields is converted to field names. If the data in the data field has a hyphen in it, this is automatically converted to an underscore when it becomes a field name.
Hyphens are legal in field names, so can we make CrossTab tolerate the string as is without changing it? If that is a breaking change, could a checkbox be added that allows users to get CrossTab to try to use the text as is and exception if the string is illegal as a field name?
Hyphens are required in the field name when using the Download tool, as some header names like "Content-Type" have hyphens in them.
The Summarize tool should have an option to ignore warnings like this:
Group Bys on Double or Float are not recommended due to rounding error.
It would be great if Alteryx could introduce a tool implementing a Decision Model Notation (DMN)-style Decision Table as an option to remove business logic from a workflow.
Arbitrary business rules are frequently implemented against datasets in Alteryx workflows before actual processing occurs. The implementation of complex business logic frequently results in a spider web of join, filter, formula, and union tools.
For example, allowing the letters in front of the column titles to be formatted differently to make them stand out on a report.
I'm from Denmark, and like several other european countries we use commas instead of dot as decimal seperator. And we use dot as thousand seperator.
So if im working in a flow with loads of price fields, lets say cost price, amount per unit, amount and amount including vat i need to do a multi field replace. Else I dont get the output i can work with in excel or other programs.
So it would be great beeing able to set seperators on a flow level, like you can in excel when importing.
Beeing able to set a date format on a flow level.
Lets say my input data is 12.12.2019 and i need 2019-12-12 in my output. If i work with several different date fields i need to use several datetime fields.
Alternate could be a multi field datetime ?
Having a search function when using the select ? And maybe a numbers order.
So if i scroll down, i could enter 3 - which means this would now be my 3. shown field?
Would be nice if in Designer customer's may want to upload and reference a " DATA DICITIONARY - METADATA REPOSITORY file when working with various input source to transform data .
Organizations that are mature in their data governance strategy implement special software that extracts, manages and provides access to data dictionary of data assets in multiple databases such as ERWIN to maintain schema for enterprise.
Within DESIGNER access to a file METADATA REPOSITORY held in DESIGNER customer may easily select a list of columns fields or attributes from that file to manipulate data elements using DESIGNER and provide all the relevant information required they wish to massage the data.
Possible Attributes that may be in data dictionary file:
I was wondering if there is a tool that could de-duplicate records after serializing (or after using Transpose Tool) with a given priority for each field in one of the keys? i.e.
Assuming for the field name NAME, the priority should be [ A, B ]
ZIP CODE -> [ C, B ]
TYPE -> [ H, D ]
The expected outcome for Id 1 should be -> JACK, 15024, PKL
Record discarded -> PETER, 15024, MID
In this case I'm using ID and Origin as keys in the Transpose Tool.
I just want to make sure there is no other route than the Python Tool.
recently loaded the new V11 and gettting used to it. one immediate gripe is the new version of the Formula Tool no longer supports multiple field actions. In the prior version I could change Data Types on many fields at once. I could move multiple fields in a block at once. there were a few other things but these are things I am sorely missing on my first use of V11. I created about 20 fields in quick succession just getting names down and then going back and putting in formula which were variations on a theme. When done I noticed the default DataType was V_WString and I wanted integer. In the past it was no big deal because I could select the block or interspersed fields and then right click to change data type for all to the same data type. it was very handy and now appears to be gone. please bring these things back.
I believe the following would help improve the functionality of Select Tool.
The idea is to have a defaulting option for each of the field in the Select Tool (which I believe should be a light weight Tool i.e. not adversely impacting performance and gives best exhaustive picture of all columns flowing through a (/particular point in ) pipeline).
Following are some of the cases where defaulting might come handy -
1) Fields which are supposed to hold monetary data - instead of Null, one can put 0.00 to help roll up summary properly.
2) Fields which are supposed to hold dates (say expiry date) - instead of Null, one can put some enterprise standards like 31-12-2099 to avoid mixing Nulls and 31-12-2099.
3) Fields which are supposed to hold purchase quantity/number of employees/number of merchandise - instead of Null, one can put 0 (and not 0.00) again to help with roll up summary.
4) Fields which are supposed to hold Currency - instead of Null, one can put USD.
5) Fields which are supposed to hold dates (say this time create date) - instead of Null, one can hardcode actual date, or an additional feature to put Now() kind of functions.
At present one of the options of achieving same might be to put a Formula Tool and to code whatever is desired inside the Formula Tool.
Benefits of having the functionality inside Select Tool would be -
1) It would be more user friendly and call for faster build to just write '0.00' or 'USD' or '31-12-2099' as compared to writing IF IsNull()... statements.
2) Inside Formula Tool, user needs to pull desired fields from the drop down and hence exhaustive view of all fields passing through pipeline is not available.
Pain in selection of fields from drop down and writing actual formulas might be aggravated with the number of columns increasing and might be more prone to human omission related errors.
I was just thinking... they might not need to fully build out a python ide, but could still reach the same objective.
You should be able to keep a python file on its own and call it in r. By doing this, you might be able to have the json/xml handling of python with the visual/stats power of R while it being nicely bundled in your workflow. This uses base functions in r and does a good job turning a pandas dataset to an r dataframe you can move along your workflow.
You could always just use this same idea to write a file somewhere and once it's written, your workflow will continue. If you do, the code is literally 1 line in r... Anyway, let me know your thoughts! 🙂
Will this work for your organization?
The community could benefit from easier integration of splitting and applying functions to grouped data. The summarize tool is great for splitting your data and applying summary statistical functions. It would be super useful to take that block just one step further, and allow users to apply any other (aggregate) function to their grouped data instead of just the built-in functions in the summarize tool. I would envision that aggregate function either being a custom function that is a combination of existing user-specified functions within Alteryx (e.g. in the formula tool) and/or even an interface that allows you to use other Alteryx macros on the grouped data.
Apply user-defined functions, or other powerful Alteryx macros to grouped and data is a very common operation in the data analyst's daily workflows and being able to apply them without reverting to batch/iterative macros in a seamless manner would be naturally helpful.