This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
I know cache-related ideas have already been posted (cache macros; cache tools), but I would like it if cache were simply built into every tool, similar to the way it is on the Input Tool.
During workflow development development, I'll run a workflow repeatedly, and especially if there is sizeable data or an R tool involved, in can get really time consuming.
I can see where managing cache could be tricky: in a large workflow processing a lot of data, nobody would want to maintain dozens of copies of that data. But there may be ways of just monitoring changes to the workflow in order to know if something needs to be rebuilt or not: e.g. suppose I cache a Predictive Tool, and then make no changes to any tool preceeding it in the workflow... the next time I run, the engine should be able to look at "cache flags" and/or "modified flags" to determine where it should start: basically start at the "furthest along cache" that has no changes preceeding it.
It would be very helpful to have hidden/interactive labels that can be utilized for an Report Map that only appear when the user is hovering over a specific part of the map. For example they only want to see the sales numbers for California, but do want to change the map. It probably makes the most sense to have this available for the HTML Report Maps. This would allow the report map to not look too crowded with labels, but still have the labels available if the user is interested in a particular part of the map. Is this something that can eventually be added to the report map tool?
When viewing spatial data in the browse tool, the colors that show a selected feature from a non-selected one are too similar. If you are zoomed out and have lots of small features, it's nearly impossible to tell which spatial feature you have selected.
Would be a great option to give the user the ability to specify the border and/or fill color for selected features. This would really help them stand out more. The custom option would also be nice so we can choose a color that is consistent with other GIS softwares we may use.
As an example, I attached a pic where I have 3 records selected but takes some scanning to find where they are in the "map".
As I understand SFTP support is planned to be included in the next release (10.5). Is there plans to support PKI based authentication also?
This would be handy as lots of companies are moving files around with 3rd parties and sometimes internally also and to automate these processes would be very helpful. Also, some company policies would prevent using only Username/Password for authentication.
The Field Summary tool is a very useful addition for quickly creating data dictionaries and analysing data sets. However it ignores Boolean data types and seems to raise a strange Conversion Error about 'DATETIMEDIFF1: "" is not a valid DateTime' - with no indication it doesn't like Boolean field types. (Note I'm guessing this error is about the Boolean data types as there's no other indication of an issue and actual DateTime fields are making it through the tool problem free.)
Using the Field Summary tool will actually give the wrong message about the contents of files with many fields as it just ignores those of a data type it doesn't like.
The only way to get a view on all fields in the table is using the Field Info tool, which is also very useful, however it should be unnecessary to 'left join' (in the SQL sense) between Field Info and Field Summary to get a reliable overview of the file being analysed.
Therefore can the Field Summary tool be altered to at least acknowledge the existence of all data types in the file?
I think that the sample tool should have a T or F port.
Lets say I want to keep first N records but would like to stream the rest of the data (the not sampled one) somewhere else in my workflow, its possible but it would be easier to have that in the sampler.
I have a very large geospatial point dataset (~950GB) . When I do a spatial match on this dataset to a small polygon, the entire large geospatial point dataset has to be read into the tool so that the geospatial query can be performed. I suspect that the geospatial query could be significantly speed up of the geospatial data could be indexed (referenced) to a grid (or multiple grids) so that the geoquery could identify the general area of overlap, then extract the data for just that area before performing the precise geoquery. I believe Oracle used (uses) this method of storing and referencing geospatial data.
At the moment, we are not able to use input data field names and its values in Output tool, mainly in the Pre-SQL and Post-SQL statement. I see some discussions on this in the community and in many scenarios we require that. It will be great if we have this option.
As a method of deploying preprocessing and ML models it would be awesome to be able to convert a workflow to java...
models are needed to be deployed into Complex event processes or decision systems. Even for SAS there is a need to implement the datastep algorithms and procs to run in JVM.
It is possible to convert a workflow into a PMML file and then use JMML package to convert that to Java. Yet the full workflow with all preprocessing alternatives and a series of ML methods may not be captured fully.
I have several .yxdb files that I’ve been appending to daily from a SQL Server table in order to extend the length of time that data is retained.
They’re massive tables, but I may only need one or two rows.
I had hoped to decrease the time it takes to get data from them by running a query on them (or a dynamic query/input) as opposed to using a filter or joining on an existing data set which would have equal values that would produce the same result as a filter.
Essentially, the input of .yxdb would have the option of inputting the full table or a SQL query just like a data connection.