This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
I think that the sample tool should have a T or F port.
Lets say I want to keep first N records but would like to stream the rest of the data (the not sampled one) somewhere else in my workflow, its possible but it would be easier to have that in the sampler.
Given redshift prefers accepting many small files for bulk loading into redshift, it would be good to be able to have a max record limit within the s3 upload tool (similar to functionality for s3 download)
The other functionality that is useful for the s3 upload tool is ability to append file names based on datetimestamp_001, 002, 003 etc similar to current output tool
At the moment, we are not able to use input data field names and its values in Output tool, mainly in the Pre-SQL and Post-SQL statement. I see some discussions on this in the community and in many scenarios we require that. It will be great if we have this option.
As a method of deploying preprocessing and ML models it would be awesome to be able to convert a workflow to java...
models are needed to be deployed into Complex event processes or decision systems. Even for SAS there is a need to implement the datastep algorithms and procs to run in JVM.
It is possible to convert a workflow into a PMML file and then use JMML package to convert that to Java. Yet the full workflow with all preprocessing alternatives and a series of ML methods may not be captured fully.
I have several .yxdb files that I’ve been appending to daily from a SQL Server table in order to extend the length of time that data is retained.
They’re massive tables, but I may only need one or two rows.
I had hoped to decrease the time it takes to get data from them by running a query on them (or a dynamic query/input) as opposed to using a filter or joining on an existing data set which would have equal values that would produce the same result as a filter.
Essentially, the input of .yxdb would have the option of inputting the full table or a SQL query just like a data connection.
It would be great if I could save a tool and its configuration for future use. I'm sure this could be done now with a macro, and that's possibly how I'll do it.
The scenario is that we have numerous data inputs that require some finessing to import. Because the field headers are not in the first row, they aren't recognized accurately as comma delimited. But some of our fields also contain commas within the field, so the solution is to add a multi-field formula to replace things like ", Inc." with " Inc." and ", Ltd." with " Ltd."
But on occasion I find new text strings that need to be added to the formula, like ", Inc" (without the period) for instance.
I use this replacement technique with several workflows, so it would be great to change it just one time and have the other instances updated automatically.
Maybe in addition to our Favorites toolbar we could have a Custom or User toolbar where we could save tools along with their configuration.