This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
I've seen various solutions/workarounds but this seems to be a table-stakes ask? When scheduling a workflow I should be able to specify which parameter values to run with. Just two of many use case examples -
Use case #1 - a single analytic app could be scheduled multiple times, each schedule specifying a different line of business
Use case #2 - credentials, allow the user to schedule a workflow with their saved credentials. When their credentials change all they would have to do is update the schedule parameters
(If you know a better way I'm all ears, not the API thought that isn't a bad way to do it, just not super Easy)
I would like to see the enhancement of being able to schedule a start and stop time for specific jobs that need to run multiple times per day, so they don't run during the nightly backup periods of the Alteryx server. This can cause the daily backups of the Alteryx Gallery to fail. Currently SQL Server Management Studio/SQL Server has the function to able to schedule start and stop times for specific job. Currently Scheduler page on Alteryx Gallery/Server lacks this feature.
Using current version of the server - you can see that there is no OAuth managed or published API endpoint for canvas delete (screenshot 1). However this API does CLEARLY exist as you can see if you inspect what happens when you hit the delete button (screenshot 2 clearly shows the API being called - but it requires user login security token)
Please can you enable this API for OAuth - the API already exists, it just needs to be exposed with the others.
I would like to set a retry count when scheduling a job, and also to specify to duration in between retries, e.g. job fails, but then will retry to run 3 times with a 5 minute wait in between each retry.
Also would like a radio button next to all workflow results, so when you get a list of failed jobs you can click all the jobs and then hit rerun. This should be complimented with select all functionality. This will prevent me from having to go into each job and rerunning, i.e. save me loadsa clicks!
We would like a built-in process that would search or, and resolve, workflows that are stuck in the "initializing" state. These seem to happen for various reasons but communication problems between the controller and workers .. usually a socket timeout.. which appears to be most problematic. It seems that these type of errors should be expected in all but the most stable environments,
Currently, the only tool that we have to solve this problem is to restart the Alteryx Service on the controller and while this works there tendency to cause some collateral damage in workflows ...erroring or restarting from their beginning.
There may be a way to solve this without restarting the service by editing Mongo using a tool like Robo 3T but that is unproven and has its own risk.
After dealing with this issue and struggling for quiet some time we think that the best option is to implement a "clean up" DB process that will run every 5 min or so, capture a list of workflows in the "initializing" state , then compare that list to one in the next 5 min cycle and fix any workflows that appear in both lists. We think that returning any stuck workflows to the queued state would be the best Fix option.
We just don't want to continue to use Restart the Service process to solve this issue and accept the collateral damage.
When I publish a workflow, I want it to have different behaviour based on the collection from which it was executed or scheduled. Let us suppose that I create a DEV collection and a PROD collection. The workflow is shared in both. If the workflow is executed from DEV, then I will have some logic to do X (output to specific folders, etc). If from PROD, the logic will do Y. The logic will be up to me. If the collection information is captured in the MongoDB layer, then I can pull it out on the fly by joining the collections, appInfos, and AS_Queue or AS_schedules collections. Querying the DB on the fly during execution and pulling that value is trivial. Once I have it, I can use it for any sort of logic as an environment sensing variable. Can you please store the info in MongoDB?
I would like to see ability to limit the number of versions a user can save of a workflow. As a server environment ages we start to see the database become unnecessarily larger because the number of version of workflows. For servers related to DEV and QA environments as an Admin I would like to be able to say users can keep their 10-25 most recent versions of their workflows. Then anything over that will start to auto-delete just like the auto save feature in Designer.
We have a MongoDB connection string that we plug in to RuntimeSettings.xml during install. Upon startup, the Mongo connection fails because RuntimeSettings.xml expects the connecting string to be encrypted. The fix for this, is for someone to manually update the MongoDB connection via the Alteryx UI. We would like a way to encrypt the MongoDB connection string for RuntimeSettings.xml before we start Alteryx for the first time.