Disclaimer… I have never used these predictive tools, this is my first attempt at a weekly challenge.
I was able to replicate the error. The issue is not bringing in the data from the same point (selection tool). I am assuming the user was using the select tool to change the data type therefore there were two different data types coming into the Append Cluster tool.
Never used the tool before...
My solution... sort of. :)
So - always keen on hearing my computer beg for mercy under the load of a multi-GB recordset ....
Cheers! This was a fun one! I learned a lot about the predictive grouping tools.
The "cannot allocate vector" error is what R gives when it runs out of memory. The workflow is trying to hold 7531GB of data in the machine’s memory at once. Would suggest the user divides the data up in batches using a batch macro.
my deep expertise in R a quick Google search it looks like the workflow is failing due to a lack of system resources (specifically, RAM).
Two possible solutions are to:
1) Throw more resource at the problem, by running the workflow on a machine with more RAM
2) Make the workflow run more efficiently, so that it can complete with the resources you currently have available
The second option is the recommended approach.
The screenshot shows that the error is occurring in the 'Append Cluster' tool.
I found the following post on Alteryx Community (https://community.alteryx.com/t5/Alteryx-Knowledge-Base/Tool-Mastery-Append-Cluster/ta-p/194965) which states that "Because this tool applies a pre-built model to a data stream, the records being assigned clusters do not need to be fed in to the tool all at once".
Therefore, I recommend changing the workflow to contain a batch macro so that records are processed by the Append Cluster tool in smaller batches.
For further guidance I highly recommend the following essential technical guides: