This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
In my environment, installing Core Data Bundle to network drive ("prepare a network install") runs for hours or days due to network factors. If interrupted, I need to start from scratch again.
These changes would greatly improve the installation:
Maintain verbose installer log to track progress and confirm that all selected files are successfully installed.
Add a repair mode to the installer to repair interrupted installations. When run in this mode, only missing/corrupt data files will be re-installed as necessary (much faster than fresh install).
Support use of a shortcut like "CURRENT" on my network drive to the real installation directory (say "Q3_2017"). This would allow me to point to the latest release and users who "register from a network location" via "CURRENT" shortcut path will always get the latest data files. As an administrator, I have no way to know if my Alteryx users are configured to use an obsolete data set that I want to clean up from my network drive.
When I prepare a network installation, I'd like to specify an expiry date and warning message that users will see in their workflows if they access obsolete data sets. Example: "Core data set XXX is obsolete as of 12/31/2017. Register current version via \\my-network-drive\alteryx\???\DataInstall.exe...". Nice to have would be a warning date and an error date to provide grace period. This would be similar to the CASS data set expiry date.
Love the functionality to create filters on the Calgary database but it would be nice to be able to select the columns you wanted returned. There are times where you only want a couple columns but the input tool will return all columns creating a larger dataset then required. You can add a select right after the input but this is after the entire dataset has been loaded into memory. Combining the two would make the Calgary input tool behave more like a database then a standard "dumb" input source.
Could we please have the option of selecting the fields, like the join tool, within the Calgary Join tool? It is especially problematic when I simply want to assign spatial object name to a point; i.e customers within a DMA or sales territory.
This would be agreat addition within the ConsumeView Matching tool as well.
As a GIS department, we use numerous spatial datasets on a daily basis. Many of these are quite large and we are looking for ways to optimize their performance. Right now, we are forced to use an indexed folder system to increase performance, but we would like to move to Calgary databases. The problem is, that Calgary databases only hold point features which limits the number of our datasets that we can use it with. If we could spatially index line and polygon features as well, that would dramatically increase the usefulness of a Calgary database.
When workig on a module with many tools, and one of those tools is the Allocate Input tool (Experian US), whenever you try to access or configure that other tools, there is a delayed reponse. I'm assuming because the Allocate data set is so large. But, it would be nice, if possible, that when accessing/configuring other tools you did not get that delay. My workaround is to Container the Allocate tool and disable it.
Functionality to append to Calgary Databases. Currently we are unioning Calgary Database to another dataset and overwriting existing Calgary database. Ability to append would cut down on data processing.
As with Output Data tool, it would be very helpful to have this option within the Calgary Loader tool. I have a series of ordered analytic apps and if I could name the Calgary database using the "Take File/Table Name from Field" option I would be able to chain the apps and be much more efficient.
Calgary CrossCount returns 1 row for every permutation and combination of possible crosscounts for the given crosscount fields. This idea is to use the additional query criteria to limit the output of rows.
Example: State Code = 'MI' and Mosaic Type = 'A01' could be our Criteria. and State Code, Mosaic Type and Gender would be our crosscount fields.
Calgary crosscount (and append) would output (51 * 71 * 4) 14,484 rows (all states * all mosaic types * all gender values)
The desired crosscount would output (1 * 1 * 4) 4 rows (MI * A01 * 4).
This is a simplified example just to demonstrate my confusion when Alteryx responded to me that the tool is working as designed. In order to reduce the crosscount output, you must restate your query criteria in a filter post the calgary tool. If the count of rows output exceeds a thresshold, you can't even use the tool as it will generate too many rows.
By defining criteria in the Calgary Input tool, only the desired output records are generated. But by defining criteria in the Calgary crosscount tools, only the desired "combinations" have values plus it outputs all permutations and combinations of other non-qualifying records.
Our team works with a lot of in-house transaction data sets that have been put into a calgary database. It would be much easier to build apps that use the calgary input tool without having to configure html code.
A great feature would be for CASS to provide the address type as Residential or Business. Better yet, further breakdown of address type into single-family, apartment, retail, office, commercial, warehouse, etc. This would be very beneficial when analyzing address data from the Tom Tom Address Points Calgary database and can allow a end-user to filter prioritize addresses in their analysis based on the type of address.
Occasionally, the Calgary Loader tool will not write out all fields passed to it. This seems to happen after writing out a certain number of fields then later, when rerunning, adding a new output field. Very annoying because you don't know it will happen until processing is complete and you examine the result. I usually manually delete the calgary files prior to rerunning, to avoid the versioning, but it still happens.
Also, please make the versioning optional with a check box, default off.