This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Recently we had a situation where installing the data packages was expected to take over 20 hours! We do not have the ability to run a machine undisturbed for this length of time at the office, and VPN automatically times out after 12 hours. Okay, so these items are company-specific, but how nice would it be to be able to download/install the data in smaller portions so you don't have to worry about setting up shop for 8-10 hours?!
I was able to copy the DataInstall.ini and name only the portions of the data I wanted to install in each session. But then I had to separate the installs into different network folders otherwise you end up overwriting the DataInstall.exe file, which users need for each install in order to register from a network location (for each install).
Needless to say, it took SEVERAL weekends and quite a few mistakes before I was able to do it this way successfully!
Please vote for data installs in smaller portions!
When workig on a module with many tools, and one of those tools is the Allocate Input tool (Experian US), whenever you try to access or configure that other tools, there is a delayed reponse. I'm assuming because the Allocate data set is so large. But, it would be nice, if possible, that when accessing/configuring other tools you did not get that delay. My workaround is to Container the Allocate tool and disable it.
As with Output Data tool, it would be very helpful to have this option within the Calgary Loader tool. I have a series of ordered analytic apps and if I could name the Calgary database using the "Take File/Table Name from Field" option I would be able to chain the apps and be much more efficient.
Functionality to append to Calgary Databases. Currently we are unioning Calgary Database to another dataset and overwriting existing Calgary database. Ability to append would cut down on data processing.
Our team works with a lot of in-house transaction data sets that have been put into a calgary database. It would be much easier to build apps that use the calgary input tool without having to configure html code.
Calgary CrossCount returns 1 row for every permutation and combination of possible crosscounts for the given crosscount fields. This idea is to use the additional query criteria to limit the output of rows.
Example: State Code = 'MI' and Mosaic Type = 'A01' could be our Criteria. and State Code, Mosaic Type and Gender would be our crosscount fields.
Calgary crosscount (and append) would output (51 * 71 * 4) 14,484 rows (all states * all mosaic types * all gender values)
The desired crosscount would output (1 * 1 * 4) 4 rows (MI * A01 * 4).
This is a simplified example just to demonstrate my confusion when Alteryx responded to me that the tool is working as designed. In order to reduce the crosscount output, you must restate your query criteria in a filter post the calgary tool. If the count of rows output exceeds a thresshold, you can't even use the tool as it will generate too many rows.
By defining criteria in the Calgary Input tool, only the desired output records are generated. But by defining criteria in the Calgary crosscount tools, only the desired "combinations" have values plus it outputs all permutations and combinations of other non-qualifying records.
Love the functionality to create filters on the Calgary database but it would be nice to be able to select the columns you wanted returned. There are times where you only want a couple columns but the input tool will return all columns creating a larger dataset then required. You can add a select right after the input but this is after the entire dataset has been loaded into memory. Combining the two would make the Calgary input tool behave more like a database then a standard "dumb" input source.