This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
At the moment if a part of your python code takes more than 30s to run, Jupyter times out and Alteryx cancels the workflow. This makes the Python Tool unusable for anything intensive and the timeout should be removed by default or be configurable per workflow.
I've made this idea as none of the solutions in these threads feel satisfactory:
We need color coding in the SQL Editor Window for input tools. We are always having to pull our code out of there and copy it into a Teradata window so it is easier to ready/trouble shoot. This would save us some time and some hassle and would improve the Alteryx user experience. ( I think you've used a couple of my ideas already. This one is a good one too. )
I would like to have all my annotations show on top of the tools by default. I can change them one at a time, but I would like to be able to change the default setting to be "Show Annotations on the Top" always.
Some of the workflows I use have multiple inputs that can take a long time to initially load. The new cache function itself has been amazing, but there is one big drawback for me: I can't cache multiple tools at the same time. Alteryx will allow me to eventually cache all of the tools I want cached, but it will take multiple times running the file. This still saves me time in the end, but it feels a bit cumbersome to set up.
I really love how I can drag and drop a file directly onto the canvas from Windows Explorer and Alteryx knows to create an Input Data tool. But when I tried it with a folder today, hoping to see a Directory Input tool appear, it wouldn't do it. Could we have a similar functionality for automatically creating a Directory Input tool?
I feel like I must be missing something, but saw a similar suggestion for TDE outputs, so maybe this really doesn't currently exist. We sometimes add descriptions to fields we create, and some inputs come with descriptions, but we can't seem to get them into the final database using the Output tool. Can there be a checkbox to persist the metadata along with the data when writing to a database?
We have recently upgraded to 18.3 and noticed that the Filter tool expression box has lost its colour coding of expression elements. The Formula tool (which still has the 18.2 look and feel) however still has the old colour coding.
It would be great if the colour coding could be included for all tools with expression editing boxes.
Today I have some workflows which have certain steps that occur after files are output. I have these set up inside of Tool Containers so that I can easily enable/disable them as I am working if I do not want to produce output for this particular run. However, sometimes if I need to troubleshoot on a workflow that I haven't worked on for awhile, I can neglect to disable these, which can cause errors. This is usually harmless, but annoying.
Having two more options on Tool Containers could really help to improve this!
Disable When Browse Tools Disabled would be useful for any analysis/debugging steps that I only want to run when I am browsing to find data, but should not run otherwise.
Disable When Output Disabled would be really useful to ensure that these tools are turned off alongside the "Disable all tools that write output" option in Workflow-Configuration-Runtime.
This would save me a lot of unnecessary error messages and moments of panic, and would make these types of workflows easier for other users to debug without extensive notes.
A common problem with the R tool is that it outputs "False Errors" like the following: "The R.exe exit code (4294967295) indicted an error"
I call this a false error because data passes out of the R script the same as if there were no error. As such, this error can generally be ignored. In my use case, however, my R tool is embedded within an iterative macro, and the error causes the iterator to stop running.
I was able to create a workaround by moving the R tool to a separate workflow and calling it from the CReW runner macro within my iterator, effectively suppressing the error message, but this solution is a bit clumsy, requires unnecessary read/writes, and uses nonstandard macros.
Presently when mapping an Excel file to an input tool the tool only recognizes sheets it does not recognize named tables (ranges) as possible inputs. When using PowerBI to read Excel inputs I can select either sheets or named ranges as input. Alteryx input tool should do the same.
In order to debug a call to a REST API - it is often necessary to take the web call, and pop this into a web browser. Can you add a second output to a RestAPI tool (a derivative of the Download tool) that has a second output that provides the full web call that was made, including the full parameterised URL. This would make it MUCH easier to debug rest API calls.
I think the Nearest Neighbor Algorithm is one of the least used, and most powerful algorithms I know of. It allows me to connect data points with other data points that are similar. When something is unpredictable, or I simply don't have enough data, this allows me to compare one data point with its nearest neighbors.
So, last night I was at school, taking a graduate level Econ course. We were discussing various distance algorithms for a nearest neighbor algorithm. Our prof discussed one called the Mahalanobis distance. It uses some fancy matrix algebra. Essentially it allows it it to filter out the noise, and only match on distance algorithms that are truly significant. It takes into account the correlation that may exists within variables, and reduces those variables down to only one.
I use Nearest Neighbor when other things aren't working for me. When my data sets are weak, sparse, or otherwise not predictable. Sometimes I don't know that particular variables are correlated. This is a powerful algorithm that could be added into the Nearest Neighbor, to allow for matches that might not otherwise be found. And allow matches on only the variables that really matter.
I love how the new (as of 2018.3) Python tool has a Jupyter notebook in the config panel. Jupyter is great, has a lot of built-in help, and is so robust that there is really no need for an external installation of any Python IDE anywhere else. I would love to see that with the R tool as well. For now (as of 2018.3), it's much easier to develop R outside of Alteryx, e.g. in R Studio, and then copy the code in the R tool.
Therefore, this request is to implement R just like Python, using Jupyter. This would allow us to script it and see our visualizations (etc) right in our Jupyter window. It would eliminate the need to have R-Studio off on the side. Here are a couple links that may hint how to make it happen:
My users love having the ability to pick objects from a reference file in the Map tool in the Interface palette. However, usually they need to pick objects that are interspersed amongst others. The Control + Left Click works great, until they pick an incorrect object. The only option is to clear the selection and start over.
Please add something as simple as Control + Left Click on a selected object will deselect it.
The new Interactive Chart is a great tool that I've been using. A big improvement to it would be to allow a secondary axis, as well as allowing data points to show on the chart. I have attached a sample chart that was created in Excel.
The new insight tool offers some great charting abilities but it does not integrate with other reporting tools. The tool doesn't support pictures,tables or any way to pull text from the data in the workflow in. This really prevents it from being a solution to any of the my reporting needs.
Currently in Alteryx if a user uses a SELECT tool to rename a field name, and then have various different tools connected in a flow to this SELECT tool, the same field name is carried throughout.
However, what if your project kept adapting to different needs and hence the field name in the SELECT tool had to be renamed again for some reason, this would cause all the linked tools in the workflow to break, especially if the previous field name is used in these tools.
It would be great to allow the user to tell Alteryx to change all linked field names in the various linked tools to the new name or keep the old one.
I understand that Alteryx at the moment does not have this feature by default, in the case where the user renames a Field name, Alteryx does not want to automatically update all linked Field Names, especially if the renamed Field Name has already been used somewhere else in the workflow. However Alteryx should provide a warning that the renamed Field Name is a duplicate !
Matlab has a similar case wherein if a user changes the name of one variable, with a simple keyboard shortcut he can rename all other variables just simply.