This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Here at Alteryx we believe in working smart, not hard. Building out reports to highlight business-critical metrics is a pretty smart way to track goals. Customizing those reports to everyone in the department, then distributing them as attachments to individual emails? That sounds like a lot of hard work. Scheduling those reports from a refreshing data source each month so you don’t have to remake or rerun the reports yourself - that’s genius. Logging into your work computer to open up Alteryx, then having to check the scheduled results before having any peace of mind those reports were delivered without a hitch? Hard.
When publishing a workflow to Gallery or Scheduler (Designer + Desktop Automation) or when packaging a workflow for export, checking the boxes for what to include and what to exclude seems to work inconsistently as of Designer 2020.2. The workaround will tide you over until you can upgrade to 2021.2.
On Alteryx version 10.1, if you try to schedule a workflow to to run from its original location on disk and the workflow path starts with a digit, for example C:\1testpath\workflow.yxmd, the workflow will fail to run on the Scheduler. When the job fails, you will see the job at the bottom of the Results view within the Scheduler window and the job date is January 01 0001 12:00:00 AM - this indicates a corrupt job. All other combinations work as expected.
This is a known issue that is being addressed. The workaround is to rename the folder so that the path does not start with a digit, for example C:\a1testpath\. Alternatively, you can schedule the job to run a copy of the workflow stored in the scheduler DB.
Upon creating a BINGO game, I came across a technique that I thought could be useful in "real world" scenarios for users who are attempting to iterate a process and then replenishing the data after a certain amount of time.