This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
Here at Alteryx we believe in working smart, not hard. Building out reports to highlight business-critical metrics is a pretty smart way to track goals. Customizing those reports to everyone in the department, then distributing them as attachments to individual emails? That sounds like a lot of hard work. Scheduling those reports from a refreshing data source each month so you don’t have to remake or rerun the reports yourself - that’s genius. Logging into your work computer to open up Alteryx, then having to check the scheduled results before having any peace of mind those reports were delivered without a hitch? Hard.
Connecting to Google Analytics is becoming more and more popular. There are a few things you need in order to use the Google Analytics macro; a Google Account (e.g., Gmail) and authorized access to an existing Google Analytics account. This article will help you get the rest of the way.
These methods should work in most versions of Alteryx up to and including 10.1.
One of the most common issues we see from clients trying to read in a .csv file is that they are receiving an error starting “Too many fields in row x”, and because of this Alteryx can’t read in the file.
There are a couple of different ways you can resolve this error.
First, in the Input Data tool, you can tell Alteryx to treat read errors as warnings to allow the file to be read in.
This will change the Error to a Warning so that Alteryx can read in the file so you can investigate.
The other method is to read the file in with no delimiter. You can do this by changing the delimiter in the Input Data tool to \0.
Once you have your data parsed back out into its fields, you can use the Dynamic Rename tool to correct your field names, a select tool to remove the original field, and a simple Trim() function to remove the extra delimiter from your data.
This process is illustrated in the attached workflow, created in version 10.1.
If you are un lucky while rendering a map in Alteryx with a Carto base map, you may encounter the error message Error: AGG error loading font (C:\WINDOWS\fonts\C:\Program). This article explains the cause of this error, and how to resolve it.
“Unhandled Exception occurred” error is thrown when you copy and paste text (Ctrl-V) using the R Tool.
Now, to witness it happening:
Looking at the error log you will see error message below:
Default Log path - C:\ProgramData\Alteryx\ErrorLogs\AlteryxGUI
If you look at the log, you can see the error is directly related to the FIPS cryptographic algorithms. According to Wikipedia, FIPS stands for Federal Information Processing Standards and it is a “standard developed by the United States federal government for use in computer systems by non-military government agencies and government contractors”.
As of right now, “Unhandled Exception occurred” error will be thrown in the R Tool if FIPS compliance is turned on .
Our Development team is fully aware of the problem with the FIPS compliance and is planning to sort this out in the future releases. As for temporary solution, you can consider turning off the FIPS compliance, of course after checking and making sure your IT manager is okay with it, and this should resolve the “Unhandled Exception occurred” error.
Here’s how you can turn off FIPS compliance:
There are other ways to turn on/off FIPS compliance and you can find them here:
You’re creating an app that involves dates. You want the user to be able to dynamically select the dates being used in the app, though. The tools you already know may not work. A Text Box would be too messy and allow lots of room for error. Pre-defined Drop Downs and List Boxes aren’t dynamic enough. Ah ha! What about that tool that looks like a calendar? The Date tool in the Interface category provides the perfect solution!
When attempting to publish a workflow to your private Gallery which contains the Publish to Tableau Server macro, you may encounter an error similar to this:
“ Cannot find macro "PublishToTableauServer_v1.08.1\Publish to Tableau Server.yxmc"
This is typically due to one of two reasons:
The Publish to Tableau Server macro has not been installed on the Alteryx Server where the Gallery is located.
The macro has been installed but has not been installed in the correct location, or the Gallery cannot access the macro for other reasons.
Please see this link for instructions on correctly installing the Publish to Tableau Server macro on an Alteryx Server.
Even when the macro has been installed properly on the server, this issue can still occur, and if it does, here is a simple workaround to get your workflow published:
Save the workflow to the Gallery: From Designer, click on File > Save As and choose your Gallery.
Add the Tableau macro components as assets of your workflow: Click on Workflow Options > Manage workflow assets.
Select all of Publish to Tableau Server related assets (currently there are four).
Including these assets when publishing should allow the workflow to publish and run successfully in the Gallery.
This error is usually caused by two things:
The file you selected is of a different format than what is selected in the File Format drop down in the Configuration window.
A .yxdb file you are using as an input has become corrupt.
Other causes can be found here: https://community.alteryx.com/t5/Alteryx-Knowledge-Base/Error-FileID-does-not-match-in-the-FileHeader/ta-p/19870
Although extremely rare, if a .yxdb file becoming corrupt is the culprit (one cause is a network connection dropping in the middle of writing the file), you may have a quick and simple way to resolve this potentially stomach-dropping issue:
Open up Windows Explorer and navigate to the file.
Right-click on the file and choose Properties.
Click on the Previous Versions tab.
If you're lucky, there is at least one option of the file that you can revert back to you, hopefully allowing you to avert disaster. Obviously, this doesn't just work with Alteryx files, so perhaps this knowledge will come in handy with other files.
Thanks for reading!
You may have run across this error, using the html plugin predictive tools (Linear Regression, Logistic Regression, Decision Tree):
Logistic Regression: Error in searchDir(dbDir, lang) : Logistic Regression: Expecting a single string value: [type=NULL; extent=0]
In 2018.2, this can happen when you have previously had an Admin version of Designer installed, but have since uninstalled. Once you've installed the 2018.2 non-Admin version with Predictive tools, these errors will now occur.
Help is on the way! (In the form of suggestions and an upcoming stable release.) You have several options. First, you can install an Admin version of Designer concurrently - 11.8, 2018.1, 2018.2, etc.
Last ditch effort: delete registry keys. This is not recommended - only delete keys if you cannot install a current version, or cannot wait until the next stable update.
Step 0) Save your license key somewhere easy to find: Options> Manage Licenses
Step 1) Open the Registry Editor (type regedit into your windows search bar) and delete the following directory:
Now, go predict stuff! Happy Alteryx-ing.
When you’re frequently writing and rewriting data to Excel spreadsheets that you use for Excel graphs and charts, it can quickly become a hassle to make and remake your reporting objects to keep them up-to-date so you’re visualizing the most recent data. A best practice to keep the hassle out of the process exists, though! If you keep your plots isolated to their own spreadsheet, referencing cell values in another sheet used to capture your data, you can simply overwrite the source data sheet and your plots will update automatically upon launching Excel. In the example below (attached in the v10.6 workflow Dynamically Update Reporting from Excel Spreadsheets.yxzp) we’ve included the workaround to make your Excel outputs seamless.
The Auto Field Tool : a tool so easy you don’t have to do anything – just put it on your canvas and viola. Automatically optimized data types. If you’re running into data type related issues and errors in your workflows, or just looking to add some speed or reduce the occupied disk space your data is hoarding – look no further than the Preparation Tool Category ’s Auto Field Tool, which reads through all the records of an input and sets the field type to the smallest possible size relative to the data contained within the column.
Why did all of my scheduled workflows fail? What does the error “Logon Failure: unknown user name or bad password” mean?
This error is due to the credentials that are used for the scheduler are no longer valid and have most likely expired. To fix this, the Run As setting has to be updated in the scheduler machine. Someone with local admin rights to the machine should go to the System Settings (Alteryx > Options > Advanced Options > System Settings) then click Next until the “Run the Worker as a Different User” menu is selected and update those credentials.
Within an Alteryx workflow you may have instances where you have to hard code a value or only want to run workflows which have a certain field schema.
Both of these cases may cause errors in the workflow when the value or schema differ to that expected. Fortunately, there are two methods you can use to try and circumvent potential errors.
Option 1 - Conditional macros – CREW Macros
Within the CREW macros pack there will be the conditional runner macro. This macro will allow you to set off another workflow if a workflow fails or succeeds. Therefore, if you are running a list of workflows from a Directory Tool and you want all the workflows to run even if a workflow fails, you can use the conditional runner to manage and continue the flow of workflows if one of them errors.
Option 2 - Expecting a set number of records or values?
You can use a combination of the Filter Tool and the Message Tool to look for patterns in the various records or files you are reading in. An example could be you are using a batch macro to read a list of files in. In each of these files you want to ensure a number of fields are included, so you transpose the data giving you a NAME & VALUE. Based on the NAME field you can use a Filter Tool and Count Records Tool to ensure you have the correct number of fields. In the false node you can have the records which didn't meet the expected value in the filter go into a Browse Tool or Output Data Tool so you can check out that file with more detail. Those records which did have the correct field headers you can crosstab the data back and continue that file through the process.
This is just a quick two ways you can look to avoid errors in a workflow, but I encourage those community members who can think of other alternatives and use cases please post below. this topic can be highly valuable for workflow conitnuity so learning more techniques can only help :)
Sometimes when reading a delimited text file (like CSV) an error like this may appear ‘Error reading “FILEPATH\MYFILE.txt”: Too many fields in record #’
The cause of this error is that too many delimiters were located in that specified record.
For example if I had a text file with the contents below:
You see that at record 6 (assuming x and y are field headers) I have an extra comma. If I’m specifying the delimiter to be a comma, Alteryx has recognized from the first record that there should only be two columns in this dataset. Since it sees three on record 6, it will error.
If you had the opposite case where you started with three commas for each record then suddenly had two like this:
You would just get a warning like this:
'Warning: Input Data (1): Record #6: Not enough fields in record'
And the last record in the third field will just be null.
So how do we get around this? Well we could always go into the text file itself and edit the file. This could be cumbersome if you have a lot of records in the text file. Instead let’s just use Alteryx to solve our problems.
First, let’s change our Input Data Tool to bring in the file as a non-delimited text file:
Now we can attach the Text To Columns Tool to separate the fields by our specified delimiter.
Now that our fields have been brought in it’s up to you to decide how handle that extra field, we could simply delete it or keep it. If you’re still wondering about its contents we can simply use a filter tool where the Record ID will be set equal to the record number that popped up in the error.
Attached to this article is a workflow I built in 10.6 that should illustrate this example.