This is a QoL-request, and I love me some QoL-updates!
While I'm developing I often need the output of a workflow as input for the next phase of my development. For example: an API run returns job location, status, and authentication ids. I want to use these in a new workflow to start experimenting what'll work best. Because of the experimenting part, I always do this in a new workflow and not cache and continue in my main flow.
Writing a temporary output file always feels like unnescesary steps, and tbh I don't want to write a file for a step that'll be gone before it reaches production. Esp if there is sensitive information in it.
I surprisingly couldn't find this anywhere else as I know it's been discussed in person on many occasions.
Basically the Formula tool needs to be smarter in many ways, but this particular post focuses on the Data Type component.
The formula tool, should not always default to V_String as the data type when entering data or a formula into the formula tool, it should look at the data type and estimate the most likely option.
I know there are times where the logical type might not be consistent in all fields, but the Data Preview and the Function of the formula should be used to determine the most likely option.
E.G. If I type a number or a date directly into the formula tool, then Alteryx should be smart enough to change the data type from the standard V_String to Int, Double or date.
This is an extension to the ideas posted here:
I often need to create a record ID that automatically increments but grouped by a specific field. I currently do it using the Multi-Row Formula tool doing [Field-1:ID]+1 because there is no group by option in the Record ID tool.
Also, sometimes I need to start at 0 but the Multi-Row Formula tool doesn't allow this so I have to use a Formula tool right after to subtract 1.
So adding a group by option to the Record ID tool would allow the user not to use the multi-row formula to do this and to start at any value wanted.
We have 'CountDistinct' and 'Concatenate' options within Summarize tool.
But 'Concatenate' displays all the instances of value for a Grouped field, this might include lot of duplicates.
It would be great to have an option like 'ConcatDistinct'.
For example -
Group by 'Branch' and 'ConcatDistinct' Customer should result as Figure 1 instead of Figure 2 -
While this is achievable in different ways currently with a set of tools, but it gets tedious when number of fields is large from which distinct values are to be captured.
As each version of Alteryx is rolled out, it would be much easier for our users and admin team to validate the new version, if Alteryx allowed parallel installs of many different versions of the software.
So - our team is currently on 11.3 - if we could roll out 11.5 in parallel then we could very easily allow users to revert to 11.3 if there are issues, or else remove 11.3 after 2-3 weeks if no issues.
The same goes for versions which are in BETA.
This would be a huge help!
When creating a workflow I generally open a "TEMPLATE" first and then immediately save it to the "NEW WORKFLOW NAME". My template includes all my preferences that aren't set naturally within the user settings and won't get RESET by them either. It has a comment box and containers as well as logos and copyrights. It would be nice to have ready access to this feature. Maybe others have standards that they want applied to all users and their workflows too.
It would be nice if we could resize the linked calgary tables box in the calgary input tool. It only shows 2 tables by default. I would like to be able to show more at times. I can change the size of the fields and query boxes, but the linked tables box is fixed at 2 rows by default.
It would be very helpful if we could pick from a list of installed calgary datasets in the dropdown menu:
we currently have the ability to choose geocoder/drivetime/Allocate datasets which are typically stored in the .ini files, but don't currently have the ability to choose calgary datasets.
I would like to be able to add an in-house built Calgary db to the list of data sources in the pull down on the Calgary Input tool. A customer database, for example.
In the picture you can see the Kalibrate Technologies Traffic Counts and TomTom US Address Points in the pull down. I would like to add my own Calgary db to the list of choices.
Love the functionality to create filters on the Calgary database but it would be nice to be able to select the columns you wanted returned. There are times where you only want a couple columns but the input tool will return all columns creating a larger dataset then required. You can add a select right after the input but this is after the entire dataset has been loaded into memory. Combining the two would make the Calgary input tool behave more like a database then a standard "dumb" input source.
In my environment, installing Core Data Bundle to network drive ("prepare a network install") runs for hours or days due to network factors. If interrupted, I need to start from scratch again.
These changes would greatly improve the installation:
Could we please have the option of selecting the fields, like the join tool, within the Calgary Join tool? It is especially problematic when I simply want to assign spatial object name to a point; i.e customers within a DMA or sales territory.
This would be agreat addition within the ConsumeView Matching tool as well.
As a GIS department, we use numerous spatial datasets on a daily basis. Many of these are quite large and we are looking for ways to optimize their performance. Right now, we are forced to use an indexed folder system to increase performance, but we would like to move to Calgary databases. The problem is, that Calgary databases only hold point features which limits the number of our datasets that we can use it with. If we could spatially index line and polygon features as well, that would dramatically increase the usefulness of a Calgary database.
When choosing "In List" values in a CYDB input, the normal Windows functions do not work (shift+click, ctl+A, ctl+click, etc.).
When having to choose, say, 20 values, it is a big annoyance to have to click each value (20 clicks).
Have been told this is a bug so I wanted to put it on your radar for a fix.
I have not found this function or a workaround, only as "recent connections" which normally, are not saved on Virtual Machines.
This would save the time it takes to find the path/folder where Calgary DBs are saved.
if this has been proposed or fixed already, please delete this idea!
As with Output Data tool, it would be very helpful to have this option within the Calgary Loader tool. I have a series of ordered analytic apps and if I could name the Calgary database using the "Take File/Table Name from Field" option I would be able to chain the apps and be much more efficient.
Our team works with a lot of in-house transaction data sets that have been put into a calgary database. It would be much easier to build apps that use the calgary input tool without having to configure html code.
I love the new (relatively) ConsumerView Decoder Tool! I used to do it the hard way, and it was fragile.
However, one thing is still missing: the Mosaic fields (MOSAIC HOUSEHOLD and MOSAIC ZIP4) - these are output from the tool as nulls. So, not only do you not get it decoded, you have to join back to the input to get the fields back as they were.
First, at least please pass them through as they were.
But preferably, decode them to the Mosaic Segment/Group names.
I realize (or couldn't find) the source for the Mosaic segment definitions is not currently in a Calgary database, but the tool is in the Calgary group.
A great feature would be for CASS to provide the address type as Residential or Business. Better yet, further breakdown of address type into single-family, apartment, retail, office, commercial, warehouse, etc. This would be very beneficial when analyzing address data from the Tom Tom Address Points Calgary database and can allow a end-user to filter prioritize addresses in their analysis based on the type of address.
Occasionally, the Calgary Loader tool will not write out all fields passed to it. This seems to happen after writing out a certain number of fields then later, when rerunning, adding a new output field. Very annoying because you don't know it will happen until processing is complete and you examine the result. I usually manually delete the calgary files prior to rerunning, to avoid the versioning, but it still happens.
Also, please make the versioning optional with a check box, default off.