Author: Michael Barone, Data Scientist
Company: Paychex Inc
Awards Category: Best Use of Predictive
Describe the problem you needed to solve
Each month, we run two-dozen predictive models on our client base (600,000 clients). These models include various up-sell, cross-sell, retention, and credit risk models. For each model, we generally group clients into various buckets that identify how likely they are to buy a product/leave us/default on payment, etc. Getting these results into the hands of the end-users who will then make decisions is an arduous task, as there are many different end-users, and each end-user can have specific criteria they are focused on (clients in a certain zone, clients with a certain number of employees, clients in a certain industry, etc.).
Describe the working solution
I have a prototype app deployed via Alteryx Server that allows the end-user to “self-service” their modeling and client criteria needs. This is not in Production as of yet, but potentially provides great accessibility to the end-user without the need of a “go-between” (my department) to filter and distribute massive client lists.
Step 1: ETL
This results in several YXDBs that are used in the models. Not all YXDBs are used in all models. This creates a central repository for all YXDBs, from which each specific model can pull in what is needed.
Once all the YXDBs and CYDBs are created, we then run our models. Here is just one of our 24 models:
The individual model scores are stored in CYDB format, to make the app run fast (since the end-user will be querying against millions and millions of records). Client information is also stored in this format, for this same reason.
Step 2: App
Step 3: Gallery
And then they can select the various client criteria:
Once done running (takes anywhere between 10 – 30 seconds), they can download their results to CSV:
Describe the benefits you have achieved
Not having to send out two dozen lists to the end-users, and the end users not having to wait for me to send them (can get them on their own). More efficient and streamlined giving them a self-service tool.
Author: Slaven Sljivar, Vice President, Analytics
Company: SmartDrive Systems, Inc.
Awards Category: Most Time Saved
Describe the problem you needed to solve
SmartDrive’s Analytics Team, which is approaching its 9th year in its existence in our 12-year-old company, is focused on three areas: 1) customer-facing analytics, 2) analytics supporting the internal teams, and 3) analytics as it is embedded within our product. To support these activities, we rely a well-developed data warehousing and business intelligence stack that includes Tableau, R, SQL Server (for relational dimensional data warehouse) and SQL Server Analysis Services cubes.
Alteryx, which we first started using only 5 months ago (March 2016), fills in a gap in our ability to quickly integrate data. Prior to Alteryx, we relied on a combination of R scrips, SQL stored procedures and SQL Server Integration Services (SSIS) jobs to develop data integration solutions. While this approach worked for us over the years, it had several drawbacks:
One specific example is a repetitive analysis that we call “Fuel Savings Analysis” (FSA). The goal of this analysis is to evaluate how much fuel our customers (commercial vehicle fleets) saved from drivers operating their vehicles differently after SmartDrive’s video event recorders were installed in the vehicles. Because video event recorders activate in response to unsafe and abrupt maneuvers, drivers tend to avoid executing such maneuvers. These maneuvers also often lead to fuel waste. For example, harsh braking wastes more kinetic energy than gradually coasting down and using the kinetic energy (and not fuel) to overcome the rolling friction and aerodynamic drag.
We had already developed a tool that automated the FSA analysis, utilizing stored procedures, R code, custom data cubes and Tableau. However, the tool required several manual steps and needed to be run for one customer at a time. As the result, SmartDrive’s Account Management team had to make a request of the Analytics team whenever the analysis needed to be run, and the Analytics team needed to expend 2 to 3 hours of effort for each request.
In April 2016, one month after we started using Alteryx, our Marketing team asked for the analysis to be done that assessed the fuel savings for all SmartDrive customers. They were interested in including that statistics in an upcoming momentum press release. Of course, this was not achievable with the existing tool, so we thought we would try to implement the workflow in Alteryx. We were ultimately successful in being able to support this request, leading to the following paragraph being included in the April 12th, 2016 press release:
Saved customers an average of $4,903 per vehicle per year—with annual per vehicle savings of $1,878 in collision exoneration, $1,784 in collision cost reduction, and $1,240 in fuel expense
Describe the working solution
Our Alteryx workflow solution issues several queries against the data warehouse, with the primary (and the largest) query representing fuel consumption and distance driven for each customer vehicle and for each week that the vehicle was driven. This is combined with a dataset that tracks when each customer site was installed with SmartDrive, so that baseline and treatment period data can be separated. An R script that employs a decision tree (rpart) is used to group vehicles and is embedded within the workflow. The key calculation for the expected fuel consumption in the treatment period (e.g. scenario that removes the effect of SmartDrive) is calculated in Alteryx, and the resulting dataset is published on Tableau Server. We authored a Tableau workbook that implements additional calculations (e.g. % fuel savings, $ savings, etc.) and allows our Account Management team to create visuals that can be shared directly with the customer. The Alteryx workflow is scheduled to run weekly every Tuesday. In less than 30 minutes, the workflow processes the entire customer dataset, with the bulk of the time being spent waiting for the data warehouse to generate the vehicle-week extract. The entire workflow is shown in the image below.
Describe the benefits you have achieved
In this particular example, Alteryx allowed us to completely streamline a process that was already largely automated using other tools. While we could have invested more time to fully automate the existing tool, that would have involved so much effort that we have repeatedly decided to de-prioritize that work.
Now that we have a fully-streamlined process, our Account Management team is able to “pull up” the Fuel Savings Analysis visualization (“report”) on their own, with up-to-date results. Also, our Marketing team is able to report on the overall actual fuel savings realized by SmartDrive customers.
Beyond the Analytics team no longer needing to spend time and effort on running the Fuel Savings Analyses, this new capability allows our Account Management team to more consistently present the fuel savings results to our customers, particularly those that are still piloting SmartDrive. This leads to increased revenue from improved pilot conversion and also greater customer satisfaction stemming from the knowledge that their investment in SmartDrive service is generating positive financial returns.
Author: Kristin Scholer (@kscholer), Insight Manager
Company: Ansira
Awards Category: Most Time Saved
Ansira, an industry-leading marketing agency in integrated real-time customer engagement, activates big data through advanced analytics, advertising technology, programmatic media and personalized customer experiences. Ansira leverages superior marketing intelligence to build deeper, more effective relationships with consumers and the retail channel partners that engage them on the local level. Marketing intelligence is infused across all disciplines and executed through digital, direct, social, mobile, media and creative execution, marketing automation, co-op and trade promotion.
Describe the problem you needed to solve
As a data-driven advertising agency, Ansira is constantly profiling customer behavior for a variety of clients in industries such as quick service restaurants, automotive brands and large retailers. Ansira’s Analytics team heavily utilizes media and consumer research that comes from the MRI Survey of the American Consumer to create Customer Behavior Reports. This large survey provides a vast database of demographics, psychographics, media opinions and shopper behavior that give insights into the actions and behaviors of the U.S. consumer. These insights help Ansira better understand consumers for new business perspectives as well as develop strategies for existing clients.
The challenge the Analytics team faced was that these rich insights were not easy to format, interpret or analyze. The data is accessed through an online portal and exported into an Excel format that does not make the data easy to manipulate. Depending on the project requirements, it could take an analyst 4-8 hours to format the data, identify survey responses that are statistically significant, build out a report to display all the information and write up a full summary. This is not cost effective and it became clear that a better way to transform this data was needed if Ansira wanted to utilize it on a regular basis.
Describe the working solution
After using Alteryx to format unfriendly Excel output for many projects, it was clear to the Analytics team that Alteryx could also be a solution for speeding up the Customer Behavior Report process. In about two days, one team member was able to create an Alteryx workflow that did all of the Excel formatting in just three seconds (this was generally taking over an hour to do manually).
Then Alteryx was taken one step further as formula tools were integrated to identify which behaviors were statistically significant for an analysis (this was taking 1-2 hours to work through manually). Next, the process was simplified one more time by incorporating the reporting tools to create a full report of all the data needed in the form of a PDF. The report even included color coding to easily identify statistically significant behaviors. Not only did this create a beautiful report in seconds but made key behaviors easy to identify, thus taking the analysis and summary process from 2-3 hours down to 15-30 minutes.
Describe the benefits you have achieved
The process that was created in Alteryx has allowed the Ansira Analytics team to offer Customer Behavior Reports to New Business and Strategy departments that can be turned around in a day instead of a week. If a full analysis is not needed, the Analytics team can turn around just the PDF data report in as little as 15 minutes (see picture below). This allows Ansira to gain additional direction on who is the consumer they are targeting, which can be instrumental in creating a new campaign.
To make this process even easier, the Analytics team has created a request form (see picture below) that anyone at Ansira can use to identify the behaviors they are interested in seeing for their client. Once the request form is received by the Analytics team, they can do a quick data pull from the MRI online portal, update the Alteryx workflow and have a full report created in under an hour.
Ansira recently had a consumer packaged goods client where the Strategy team needed to learn more about the difference in behavior patterns between Millennials and Non-Millennials who purchased 16 specific products. The Analytics team was able to pull data from MRI on these 16 products, run it through the Customer Behavior Report workflow and create 16 individual reports for each product comparing Millennials and Non-Millennials purchase behaviors in less than 4 hours. Without Alteryx, this would have taken a single analyst almost a full week to complete and likely would have never even been a possibility due to budget and capacity constraints.
Creating these Consumer Behavior Reports have become a regular occurrence with two to three requests coming into the Analytics department each week. With the help of Alteryx, these reports have become a large asset to Ansira as they provide very impactful information without a lot of effort.
Author: Alexandra Wiegel, Tax Business Intelligence Analyst
Company: Comcast Corp
Awards Category: Best Business ROI
A Corporate Tax Department is not typically associated with a Business Intelligence team sleekly manipulating and mining large data sources for insights. Alteryx has allowed our Tax Business Intelligence team to provide incredibly useful insight to several branches of our larger Tax Department. Today, almost all of our data is in Excel or csv format and so data organization, manipulation and analysis have previously been accomplished within the confines of Excel, with the occasional Tableau for visualization. Alteryx has given us the ability to analyze, organize, and manipulate very large amounts of data from multiple sources. Alteryx is exactly what we need to solve our colleague’s problems.
Describe the problem you needed to solve
Several weeks ago we were approached about using Alteryx to do a discovery project that would hopefully provide our colleagues further insight into the application of tax codes to customer bills. Currently, our Sales Tax Team uses two different methods to apply taxes to two of our main products respectively. The first method is to apply Tax Codes to customer bill records and then run those codes through software that generates and applies taxes to each record. The second method is more home-grown and appears to be leading to less consistent taxability on this side of our business.
Given that we sell services across the entire country, we wanted to explore standardization across all our markets. So, our Sales Tax team tasked us with creating a workflow that would compare the two different methods and develop a plan towards the goal of standardization and the effect it would have on every customer’s bills.
Describe the working solution
Our original source file was a customer level report where the records were each item (products, fees, taxes, etc.) on a customer’s bill for every customer in a given location. As it goes with data projects, our first task was to cleanse, organize, and append the data to make it uniform.
The next step was to add in the data from several data sources that we would ultimately need in order to show the different buckets of customers according to the monetary changes of their bills. Since these sources were all formatted differently and there was often no unique identifier we could use to join new data sources to our original report. Hence, we had to create a method to ensure we did not create duplicate records when using the join function. We ended up using this process multiple times (pictured below)
And so, the workflow followed. We added tax descriptions, new codes, and other information. We added calculated fields to determine the amount of tax that should be owed by each customer today, based on our current coding methods.
After we had layered in all the extra data that we would need to create our buckets, we distinguished between the two lines of business and add in the logic to determine which codes, at present, are taxable.
For the side of our business whose taxability is determine by software, you will notice that the logic is relatively simple. We added in our tax codes using the same joining method as we did above and then used a single join to a table that lists the taxable codes.
For the side of our business whose taxability is determine by using our home-grown method, you can see below that the logic is more complicated. Currently, the tax codes for this line of business are listed in such a way that requires us to parse a field and stack the resulting records in order to isolate individual codes. Once we have done this we can then apply the taxability portion. We then have to use this as a lookup for the actual record in order to determine if a record contains within the code column a tax code that has been marked as taxable. Or in other words, to apply our home-grown taxability logic is complicated, time consuming, and leaves much room for error.
Once we stacked all this data back together we joined it with the new tax code table. This will give us the new codes so that the software can be used for both lines of business. Once we know these new codes, we can simulate the process of the software and determine which of the new codes will be taxable.
Knowing whether or not codes are taxable helps us hypothesize about how problematic a geographic location may end up being for our team, but it does not tell us the dollar amount of taxes that will be changing. To know this we must output files that will be run through the real software.
Hence, once we have completed the above data manipulation, cleansing, and organization, we extract the data that we want to have run through the software and reformat the records to match the necessary format for the software recognition.
We created the above two macros to reformat the columns in order to simply this extensive workflow. Pictured below is the top macro. The difference between the two resides in the first select tool where we have specified different fields to be output.
After the reformatting, we output the files and send them to the software team.
When the data is returned to us, we will be able to determine the current amount of tax that is being charged to each customer as well the amount that will be charged once the codes are remapped. The difference between these two will then become our buckets of customers and our Vice President can begin to understand how the code changes will affect our customer’s bills.
Describe the benefits you have achieved
Although this project took several weeks to build in Alteryx, it was well worth the time invested as we will be able to utilize it for any other locations. We have gained incredible efficiency in acquiring insight on this standardization project using Alteryx. Another benefit we have seen in Alteryx is the flexibility to make minor changes to our workflow which has helped us easily customize for different locations. All of the various Alteryx tools have made it possible for the Tax Business Intelligence team to assist the Tax Department in accomplishing large data discovery projects such as this.
Further, we have begun creating an Alteryx app that can be run by anyone in our Tax Department. This frees up the Tax Business Intelligence team to work on other important projects that are high priority.
A common benefit theme amongst Alteryx users is that Alteryx workflows save companies large amounts of time in data manipulation and organization. Moreover, Alteryx has made it possible (where it is impossible in Excel) to handle large and complicated amounts of data and in a very user friendly environment. Alteryx will continue to be a very valuable tool which the Tax Business Intelligence team will use to help transform the Tax department into a more efficient, more powerful, and more unified organization in the coming years.
How much time has your organization saved by using Alteryx workflows?
We could never have done this data discovery project without using Alteryx. It was impossible to create any process within Excel given the quantity and complexity of the data.
In other projects, we are able to replicate Excel reconciliation processes that are run annually, quarterly, and monthly in Alteryx. The Alteryx workflows have saved our Tax Department weeks of manual Excel pivot table work. Time savings on individual projects can range from a few hours to several weeks.
What has this time savings allowed you to do?
The time savings has been invaluable. The Tax Department staff are now able to free themselves of the repetitive tasks in Excel, obtain more accurate results and spend time doing analysis and understanding the results of the data. The “smarter” time spent to do analyses will help transform the Tax Department with greater opportunities to further add value to the company.
Author: Cesar Robles, Sales Intelligence Manager
Company: Bavaria S.A.
Awards Category: Best Business ROI
Describe the problem you needed to solve
In September 30th 2015, a gossip widespread through whatsapp reduces our Pony Malta sales to 40% of normal levels. The social networks’ gossip that impacts a brand destroys brand equity and creates distrust in our customers. Our company executes a 1st stage plan that helps to spread the gossip in the first weeks to more customers increasing the crisis. In Colombia no brand had suffered an attack like this before.
Describe the working solution
The Alteryx solution was develop to design and decision tree that define which customers has an relevant impact in sales volume in 5 groups that allows define differentiated protocols to recover our sales in a healthy way. These 5 groups were:
Citizens: Actual Customers without any impact related to social network crisis.
Refugees: Customers that reduce significantly (<50%) his rate of sales related to social network crisis.
Deportees: Customers that didn’t bought our brand related to social network crisis.
Pilgrims: Customers with doubts about our products related to social network crisis.
Aliens: New customers without any impact related to social network crisis.
Our gap in crisis was 180k PoS (Point of Sales) impacting 92 KHl (Kilo-hecto-liters)
This workflow runs monthly and uses multiple sources of information in SQL server related to Customer properties and historic sales levels. We report some results in Excel and Java applications to define our performance in recovery actions. Actually we are migrating to in database process to optimize the algorithm performance and use Tableau to manage our visualization process.
Figure 1. Decision Tree description
Figure 2. 1st Quarter Deportees results
Figure 3. 1st Quarter Refugees results
Figure 4. 1st Quarter Citizens results
Figure 5. Numerical Distribution Initial and End State
Figure 6. Blending Workflow
Figure 7. Decision Tree workflow
Figure 8. Hierarchy and Priority workflow
Describe the benefits you have achieved
The project defines a new way to customer segmentation in our company. We use the same algorithm to define not only crisis contingence, also we used to brand expansion and price control process including geographical variables and external info of our providers (Nielsen, YanHass, Millward Brown).
The solution had not been implemented before Alteryx. An estimated time saving show us that initial state needs 2 or 3 weeks to develop compared with 4 or 5 days that we used in Alteryx (We just used it 1 month ago in the firs solution). Right now our response time is less than 2 days in similar solutions.
In Business terms, we achieve to recover 100k PoS (approximately 25% of all Colombia Market) and increase our sales in 75% of normal levels in the first 3 months. In August 2016, we recover our normal levels of sales with the trade marketing actions focused support by Alteryx workflow.
Author: Jennifer Jensen, Sr. Analyst and team members Inna Meerovich, RJ Summers
Company: mcgarrybowen
mcgarrybowen is a creative advertising agency that is in the transformation business. From the beginning, mcgarrybowen was built differently, on the simple premise that clients deserve better. So we built a company committed to delivering just that. A company that believes, with every fiber of its being, that it exists to serve clients, build brands, and grow businesses.
Awards Category: Best Business ROI
Describe the problem you needed to solve
Mcgarrybowen creates hundreds of pieces of social creative per year for Fortune 500 CPG and Healthcare brands, on platforms including Facebook and Twitter. The social media landscape is constantly evolving especially with the introduction of video, a governing mobile-first mindset, and interactive ad units like carousels, but yet the capabilities for measuring performance on the platforms have not followed as closely.
Our clients constantly want to know, what creative is the most effective, drives the highest engagement rates, and the most efficient delivery? What time of day, day of week is best for posting content? What copy and creative works best? On other brands you manage, what learnings have you had?
But, therein lies the challenge. Answers to these questions aren’t readily available in the platforms, which export Post-Level data in raw spreadsheets with many tabs of information. Both Facebook and Twitter can only export 90 days of data at a time. So, to look at client performance over longer periods of time and compared to their respective categories, and derive performance insights that drive cyclical improvements in creative – we turned to Alteryx.
Describe the working solution
Our Marketing Science team first created Alteryx workflows that blended multiple quarters and spreadsheet tabs of social data for each individual client. The goal was to take many files over several years that each contained many tabs of information, and organize it onto one single spreadsheet so that it was easily visualized and manipulated within Excel and Tableau for client-level understanding. In Alteryx, it is easy to filter out all of the unnecessary data in order to focus on the KPIs that will help drive the success of the campaigns. We used “Post ID,” or each post’s unique identifying number, as a unifier for all of the data coming in from all tabs, so all data associated with a single Facebook post was organized onto a single row. After all of the inputs, the data was then able to be exported onto a single tab within Excel.
After each client’s data was cleansed and placed into a single Excel file, another workflow was made that combined every client’s individual data export into a master file that contained all data for all brands. From this, we can easily track performance over time, create client and vertical-specific benchmarks, and report on data efficiently and effectively.
Single Client Workflow
Multi-Client Workflow
Describe the benefits you have achieved
Without Alteryx, it would take countless hours to manually work with the social data in 90 day increments and manipulate the data within Excel to mimic what the Alteryx workflow export does in seconds. With all of the saved time, we are able to spend more time on the analysis of these social campaigns. Since we are able to put more time into thoughtful analysis, client satisfaction with deeper learnings has grown exponentially. Not only do we report out on past performance, but we can look toward the future and more real-time information to better analyze and optimize.
Author: Brett Herman ( @brett_hermann ) , Project Manager, Data Visualization
Company: Cineplex
Cineplex Inc. (“Cineplex”) is one of Canada’s leading entertainment companies and operates one of the most modern and fully digitized motion picture theatre circuits in the world. A top-tier Canadian brand, Cineplex operates numerous businesses including theatrical exhibition, food service, amusement gaming, alternative programming (Cineplex Events), Cineplex Media, Cineplex Digital Media, and the online sale of home entertainment content through CineplexStore.com and on apps embedded in various electronic devices. Cineplex is also a joint venture partner in SCENE – Canada’s largest entertainment loyalty program.
Awards Category: Most Time Saved
Describe the problem you needed to solve
Incremental/Uplift Modelling is a popular method of evaluating the success of business initiatives at Cineplex. Its effectiveness at measuring the change in consumer behavior over time creates a high demand to produce this kind of analysis for various departments in the organization. Due to the large amount of requests we receive, the ‘Incremental Lift Model’ was developed to take in user-defined inputs, and output the results within a short period of time.
Describe the working solution
Our solution works through a four step process. The first step is for the client to populate the ‘study input form’ in order to define their study parameters and the type of study they want to run.
Visual 1: Study Input Form
The second step is to update/materialize our loyalty data that’s inputted into the model (yxdb format). We do this so that the model doesn’t put stress on our SQL Server databases, and to increase the model’s performance.
Visual 2: Update/Materialize Alteryx Input Data
The third step is the core of the incremental lift modelling. A macro processes one study at a time by pointing to the user defined inputs made in the first step.
Visual 3: Study Numbers are selected and passed through the incremental lift macro, and saves the output to SQL.
The data will then be passed through one of several macros depending on the study type, and filtered down based on the inputs defined by the user in the study input form. All data sources are joined together and lift calculations are made, which are then outputted into a master SQL Table ready to be visualized.
Visual 4: Incremental Lift Modelling based on study type selected.
The results are visualized using a Tableau Dashboard in order to share and communicate the results of the study back to the business clients.
Visual 5: Tableau Dashboard to explain to the business how the incremental lift model makes its calculations.
Describe the benefits you have achieved
The overarching goal of this project was twofold; to minimize the amount of work required to process business requests while maximizing the output generated, and to develop a means of delivering the results in a consistent manner. Both of these goals contribute greatly to our ROI by virtually eliminating all time spent executing requests that come in, and by minimizing time spent meeting with business users to explain how the incremental lift model works and how to interpret the results.
Author: Mandy Luo, Chief Actuary and Head of Data Analytics
Company: ReMark International
Awards Category: Best Use of Predictive
As a trained Statistician, I understand why "70% data, 30% model" is not an exaggeration. Therefore, before applying any regression models, I always make sure that input data are fully reviewed and understood. I use various data preparation tools to explore, filter, select, sample or join up data sources. I also utilize the data investigation tools to conduct or validate any statistical evaluation. Next, I would usually choose 3-5 predictive modeling candidates depending on the modeling objective and data size. I often include one machine learning methods in order to at least benchmark other models. After the modeling candidates finish running, I would select the best model based on both art (whether the coefficients look reasonable based on my understanding of the data and business) and science (statistical criteria's like the goodness of fit, P-value and cumulative lift etc.). I am also often using the render function for model presentation and scoring/sorting function for model validation and application.
Describe the problem you needed to solve
ReMark is not only an early adopter in predictive modeling for life insurance, but also a true action taker on customer centricity by focusing on customer lifetime analytics (instead of focusing on 'buying' only). In this context, we need to 'join up' our predictive models on customer response, conversion and lapse in order to understand the most powerful predictors that drive customer activities across pre and post sales cycle. We believe the industry understand that it is insufficient to only focus on any single customer activity, but is still exploring how this can be improved through modeling and analytics, which is where we can add value.
Describe the working solution
Our working solution goes with the following steps:
Describe the benefits you have achieved
Author: Sintyadi Thong ( @MizunashiSinayu ), Pre-Sales Consultant, Karen Kamira & Harry Yusuf
Company: Lightstream Analytics PTE Ltd.
LightStream Analytics focuses on performance management, big data, advanced analytics, and innovative visualization through cloud SaaS applications, mobile, and traditional on-site systems. LightStream Analytics is well-positioned to deliver the most advanced products and services by capitalizing on its significant regional presence in Singapore and Indonesia. The combined offices have over 60 employees with deep technical and senior business experience. The company leverages our existing technical support and R&D centers in Indonesia and China to develop solutions which disrupt customary methods of data analysis and give clients access to revolutionary tools for understanding their data. LightStream Analytics has partnered with more than 100 multinational and local clients to integrate, structure, analyze, and visualize information to measure their business performance and drive enterprise value growth.
Awards Category: Most Time Saved
Describe the problem you needed to solve
One of our clients tried to implement one of the most Business Intelligence solution to help them grow their business through another company (we can pretty much say our competitor). However, there is one thing which hinders their development of the BI. When usually most companies want to see the date of sales (on which date their agents perform sales), this company would like to see the other way around, they would like to see ‘on which dates their agents do not perform sales activity’. For them this is very important. The BI developers hit a dead-end on this thing, and therefore I came with Alteryx.
Describe the working solution
The BI I previously mentioned is QlikView. Well Qlik can do it, and I can guarantee. But it involves heavy scripting and logic. With heavy scripting it will also mean it requires heavy resources to perform the run (will only be visible when running with low RAM). Alteryx on the other hand can do this easily using drag-and-drop and repeatable workflow, so I feed Alteryx with the actual sales data, perform several left joins, filter, and unique. Alteryx requires no scripting, and to be honest I am not even an IT guy, I know nothing about SQL and programming, but I can create this workflow easily. So we proposed to have Alteryx prepare and blend the data before feeding it to QlikView, therefore it will help to make the data visible before feeding it to QlikView and lessens the burden on QlikView. While the client has not yet confirmed whether they will get Alteryx or not, it is really satisfying and rewarding to solve this problem easily while others had hardships in getting this result.
Describe the benefits you have achieved
While I created the workflow in only an hour vs their 2 weeks development for this one case (and in which they failed after 2 weeks), this shows how much of a time savings the client would get if they developed QlikView alongside with Alteryx. Alteryx will help the customers to get results faster and perform an advanced ETL which might be hard to do in traditional SQL language.
Author: Rana Dalbah, Director - Workforce Intelligence & Processes
Company: BAE Systems
Awards Category: Most Unexpected Insight - Best Use Case for Alteryx in Human Resources
Working in Human Resources, people do not expect us to be technology savvy, let alone become technology leaders and host a "Technology Day" to show HR and other functions the type of technology that we are leveraging and how it has allowed us, as a team, to become more efficient and scalable.
Within the Workforce Intelligence team, a team responsible for HR metrics and analytics, we have been able to leverage Alteryx in a way that has allowed us to become more scalable and not "live in the data", spending the majority of our time formatting, cleansing, and re-indexing. For example, Alteryx replaced both Microsoft T-SQL and coding in R for our HR Dashboard, which allowed us to decrease the pre-processing time of our HR Dashboard from 8-16 hours per month to less than 10 minutes per month, which does not account for the elimination of human intervention and error.
With the time savings due to Alteryx, it has allowed us to create custom metrics in the dashboard at a faster rate to meet customer demands. In addition, it has also given us the opportunity to pursue other aspects of Alteryx forecast modeling, statistical analysis, predictive analytics, etc. The fact that we are able to turn an HR Dashboard around from one week to two days has been a game changer.
The HR dashboard is considered to have relevant data that is constantly being used for our Quarterly Business Reviews and has attracted the attention of our CEO and the Senior Leadership. Another use that we have found for Alteryx is to create a workflow for our Affirmative Action data processing. Our Affirmative Action process has lacked consistency over the years and has changed hands countless times, with no one person owning it for more than a year. After seeing the capabilities for our HR Dashboard, we decided to leverage Alteryx to create a workflow for our Affirmative Action processing that took 40 hours of work down to 7 minutes with an additional hour that allows for source data recognition
recognition and correction. We not only have been able to cut down a two or three month process to a few minutes, but we also now have a documented workflow that lists all the rules and exceptions for our process and would only need to be tweaked slightly as requirements change.
For our first foray into predictive analytics, we ran a flight risk model on a certain critical population. Before Alteryx, the team used SPSS and R for the statistical analysis and created a Microsoft Access database to combine and process at least 30 data files. The team was able to run the process, with predictive measures, in about 6 months. After the purchase of Alteryx, the workflow was later created and refined in Alteryx, and we were able to run a small flight risks analysis on another subset of our population that took about a month with better visualizations than what R had to offer. By reducing the data wrangling time, we are able to create models in a more timely fashion and the results are still relevant.
The biggest benefit of these time-savings is that it has freed up our analytics personnel to focus less on “data chores” and more on developing deeper analytics and making analytics more relevant to our executive leadership and our organization as a whole. We’ve already become more proactive and more strategic now that we aren’t focusing our time on the data prep. The combination of Alteryx with Tableau is transformative for our HR, Compensation, EEO-1, and Affirmative Action analytics. Now that we are no longer spending countless hours prepping data, we’re assisting other areas, including Benefits, Ethics, Safety and Health, Facilities, and even our Production Engineering teams with ad-hoc analytics processing.
Describe the problem you needed to solve
A few years ago, HR metrics was a somewhat foreign concept for our Senior Leadership. We could barely get consensus on the definition of headcount and attrition. But in order for HR to bring to the table what Finance and Business Development do: metrics, data, measurements, etc. we needed to find a way to start displaying relevant HR metrics that can steer Senior Leadership in the right direction when making decisions for the workforce. So, even though we launched with an HR Dashboard in January of 2014, it was simple and met minimum requirements, but it was a start. We used Adobe, Apache code and SharePoint, along with data in excel files, to create simple metrics and visuals. In April 2015, we launched the HR Dashboard using Tableau with the help of a third party that used Microsoft SQL server to read the data and visualize it based on our requirements. However, this was not the best solution for us because we were not able to make dynamic changes to the dashboard in a fast timeframe. The dashboard was being released about two weeks after fiscal month end, which is an eternity in terms of relevance to our Senior Leadership.
Once we had the talent in-house, we were able to leverage our technological expertise in Tableau and then, with the introduction of Alteryx, create our workflows that cut down a 2 week process into a few days - including data validation and dashboard distribution to the HR Business Partners and Senior Leadership. But why stop there? We viewed Alteryx as a way to help refine existing manual processes: marrying multiple excel files using vlookups, pivot tables, etc. that were not necessarily documented by the users and cut back on processing time. If we can build it once and spend minimal time maintaining the workflow, why not build it? This way, all one has to do in the future is append or replace a file and hit the start button, and the output is created. Easy peasy! That is when we decided we can leverage this tool for our compliance team and build out the Affirmative Action process, described above, along with the EE0-1 and Vets processing.
What took months and multiple resources now takes minutes and only one resource.
Describe the working solution
The majority of the data we are using comes from our HCM (Human Capital Management Database) in excel based files. In addition to the HCM files, we are also using files from our applicant tracking system (ATS), IMPACT Awards data, Benefit provider, 401K, Safety and Health data, and pension providers.
Anything that does not come out of our HCM system are coming from a third party vendor. These files are used specifically for our HR dashboard, Affirmative Action Plan workflow, Safety & Health Dashboard, and our benefits dashboard.
In addition to dashboards, we have been able to leverage the mentioned files along with survey data and macro-economic factors for our flight risk model. We have also leveraged Google map data to calculate the commute time from an employee's home zip code to their work location zip code. This was a more accurate measurement of time spent on the road to and from work when compared to distance.
The ultimate outputs vary: an HR dashboard that tracks metrics such as demographics, headcount, attrition, employee churn/movement, rewards and exit surveys is published as a Tableau workbook. The Flight Risk analysis that allows us to determine what factors most contribute to certain populations leaving the company; a compensation dashboard that provided executives a quick way to do merit and Incentive Compensation planning includes base pay, pay ratios, etc. is also published as a Tableau Workbook.
This workflow has as its input our employee roster file, which includes the employee’s work location and supervisor identifiers and work locations going up to their fourth level supervisor. For the first step of processing, we used stacked-joins to establish employee’s supervisor hierarchies up to the 8th level supervisor. We then needed to assign initial “starting location” for an employee based on the location type. That meant “rolling up” the employee’s location until we hit an actual company, not client, site. We did this because Affirmative Action reporting requires using actual company sites. The roll-up was accomplished using nested filters, which is easier to see, understand, modify, and track than a large ELSEIF function (important for team sharing).
Once the initial location rollup was completed, we then needed to rollup employees until every employee was at a site with at least 40 employees. While simply rolling all employees up at once would be quick, it would also result in fewer locations and many employees being rolled up too far from their current site which would undermine the validity and effectiveness of our Affirmative Action plan. Instead, we used a slow-rolling aggregate sort technique, where lone employees are rolled up into groups of two, then groups of two are rolled up into larger groups, and so on until sites are determined with a minimum of 40 employees (or whatever number is input). The goal is to aggregate employees effectively, while minimizing the “distance” of the employee from their initial site. This sorting was accomplished using custom-built macros with a group size control input that can be quickly changed by anyone using the workflow.
The end result was the roster of employees with the original data, with new fields identifying their roll-up location, and what level of roll-up from their initial location was needed. A small offshoot of “error” population (usually due to missing or incorrect data) is put into a separate file for later iterative correction.
Previously, this process was done through trial and error via Access, and Excel. That process, was not only much slower and more painstaking, but it also tended to result in larger “distances” of employees from initial sites then was necessary. As a result, our new process is quicker, less error-prone, and arguably more defensible than its predecessor.
One of the Macros used in AAP:
Describe the benefits you have achieved
Alteryx has enabled our relatively small analytics shop (3 people) to build a powerful, flexible and scalable analytics infrastructure without working through our IT support. We are independent and thus can reply to the user's custom requests in a timely fashion. We are seen as agile and responsive - creating forecasting workflows in a few days to preview to our CEO and CHRO instead of creating Power Point slides to preview for them a concept. This way, we can show them what we expect it to look like and how it will work and any feedback they give us, we can work at the same time to meet their requirements. The possibilities of Alteryx, in our eyes, are endless and for a minimal investment, we are constantly "wowing" our customers with the service and products we are providing them. In the end, we have been successful in showing that HR can leverage the latest technologies to become more responsive to business needs without the need for IT or developer involvement.
Author: Katie Snyder, Marketing Analyst
Company: SIGMA Marketing Insights
Awards Category: Most Time Saved
We've taken a wholly manual process that took 2 hours per campaign and required a database developer, to a process that takes five minutes per campaign, and can be done by an account coordinator. This frees our database developers to work on other projects, and drastically reduces time from data receipt to report generation.
Describe the problem you needed to solve
We process activity files for hundreds of email campaigns for one client alone. The files come in from a number of different external vendors, are never in the same format with the same field names, and never include consistent activity types (bounces or opt-outs might be missing from one campaign, but present in another). We needed an easy, user-friendly way for these files to be loaded in a consistent manner. We also needed to add some campaign ID fields that the end user wouldn't necessarily know - they would only know the campaign name.
Describe the working solution
Using interface tools, we created an analytic app that allowed maximum flexibility in this file processing. Using a database query and interface tools, Alteryx displays a list of campaign names that the end user selects. The accompanying campaign ID fields are passed downstream. For each activity type (sent, delivered, bounce, etc), the end user selects a file, and then a drop down will display the names of all fields in the file, allowing the user to designate which field is email, which is ID, etc. Because we don't receive each type of activity every time, detours are placed to allow the analytic app user to check a box indicating a file is not present, and the workflow runs without requiring that data source.
All in all, up to six separate Excel or CSV files are combined together with information already existing in a database, and a production table is created to store the information. The app also generates a QC report that includes counts, campaign information, and row samples that is sent to the account manager. This increases accountability and oversight, and ensures all members of the team are kept informed of campaign processing.
Process Opt Out File - With Detour:
Join All Files, Suppress Duplicates, Insert to Tables:
Generate QC Report:
Workflow Overview:
QC Report Example:
Describe the benefits you have achieved
In making this process quicker and easier to access, we save almost two hours of database developer time per campaign, which accounts for at least 100 hours over the course of the year. The app can be used by account support staff who don't have coding knowledge or even account staff of different accounts without any client specific knowledge, also saving resources. Furthermore, the app can be easily adapted for other clients, increasing time savings across our organization. Our developers are able to spend time doing far more complex work rather than routine coding, and because the process is automated, saves any potential rework time that would occur from coding mistakes. And the client is thrilled because it takes us less time to generate campaign reporting.
Author: Alberto Guisande (@Aguisande), Services Director
Awards Category: Most Unexpected Insight - Proving teachers wrong - Apples & Oranges can be compared! (thanks to Alteryx)
Describe the problem you needed to solve
Our customer is a Public Transportation company, in charge of buses going around the city of Panama. They transport more than 500K passengers a day (1/6 of the total population of the country). Almost 400 routes, with 1,400 buses going around the city all days, working 24/7, reporting position every a few seconds. The company is supporting its operation with a variety of tools, but at the time to put all data together, they realized there was no "point of contact" in the data. They have to compare apples & oranges! Really? Why does the saying exist? Because you can't! So we started trying to do the impossible!
BTW, the business questions are pretty simple (once you got the data!): What route was every bus in, when every transaction occurred? What is the demand of every route? and for every stop?
Describe the working solution
Working with Alteryx, we were able to analyze data coming from three different sources, where the only common information was some LATITUDE & LONGITUDE (taken with different equipment, so the accuracy was, at least, questionable) at some random points in time. The data was received in several files:
So far, the easy part! We managed very easily to get all this info together. Now the tricky part: There mainly two operational datasets: AVL (Every position of every bus, every n seconds, where n is an arbitrary number between 0 and what the piece of hardware wanted to use). BTW, a huge amount of data every day.
Transactions: transactions registered in time, in a bus. As you may infer, there are no data in common that allow us to match records beside an arbitrary range of latitude and longitude in some random time ranges. Because of how everything is reported, the bus may be passing in front a stop that is part of another route, or stopping far from the designated stop.
Describe the benefits you have achieved
With this solution, the company can start analyzing activity per route, demand per bus, route, stop, etc. Without Alteryx, this customer information still be looking like apples and oranges! We were able to make it sense and allow them to use it to get insights.
Colorful note(and some ego elevator) : 5 other vendors took the challenge. No other one could reach a glimpse of solution (of course, "no Alteryx, no gain").
Author: Scott Elliott (@scott_elliott) , Senior Consultant
Company: Webranz Ltd
Awards Category: Best Use of Alteryx Server
We are using the server to store Alteryx Apps that get called by the "service bus" and perform calculations and write the results into a warehouse where growers can log into a web portal and check the results of the sample.
Describe the problem you needed to solve
Agfirst BOP is a agricultural testing laboratory business that perform scientific measurement on Kiwifruit samples it receives from 2500 growers around New Zealand. In peak season it tests up to 1000 samples of 90 fruit per day. The sample test results trigger picking of the crop, cool storage, shipping and sales to foreign markets. From the test laboratory the grower receives notification of the sample testing being completed. They log into a portal to check the results. Agfirst BOP were looking for a new technology to transform the results from the service bus up to the web portal which gave them agility around modifying or adding tests.
Describe the working solution
We take sample measurement results from capture devices. These get shipped to a landing warehouse. There is a trigger which calls the Alteryx Application residing on the Alteryx server for each sample and test type. The Alteryx App then performs a series of calculations and publishes the results into the results warehouse. The grower is now able to login to the web portal and check their sample. Each App contains multiple batch macros which allow processing sample by sample. Some of the tests have a requirement for the use of advanced analytics. These tests call R as part of the App. The use of macros is great as it provide amazing flexibilty and agility to plug in or plug out new tests or calculations. Having it on Alteryx Server allows it to be enterprise class by giving it the ability to be scaled and flexible at the same time. As well as being fully supported by the infrastructure team as it is managed within the data centre rather than on a local desktop.
App:
Batch Macro:
Describe the benefits you have acheived
The benefits realised include greater agility around adding/removing sample tests via the use of Macros. We are able to performed advanced analytics by calling R and it futures proofs the business by enabling them to choose any number of vendors and not be limited by the technology because of the ability of Alteryx to blend multiple sources. It gives them amazing flexibility around future technology choices and it is all supported and backed up by the infrastructure team because it sits within the datacentre and they have great comfort in knowing it's not something sitting under someones desk.
Author: Jack Morgan (@jack_morgan), Project Management & Business Intelligence
Awards Category: Most Time Saved
After adding up the time savings for our largest projects we came up with an annual savings of 7,736 hours - yea, per year! In that time, you could run 1,700 marathons, fill 309,000 gas tanks or watch 3,868 movies!! Whaaaaaaaaaaaaat! In said time savings, we have not done any of the previously listed events. Instead, we've leveraged this time to take advantage of our otherwise unrealized potential for more diverse projects and support of departments in need of more efficiency. Other users that were previously responsible for running these processes now work on optimizing other items that are long overdue and adding value in other places by acting as project managers for other requests.
Describe the problem you needed to solve
The old saying goes, Time is of the essence, and there are no exceptions here! More holistically, we brought Alteryx into our group to better navigate disparate data and build one-time workflows to create processes that are sustainable and provide a heightened level of accuracy. In a constraint driven environment my team is continuously looking for how to do things better. Whether that is faster, more accurately or with less needed oversight is up to our team. The bottom line is that Alteryx provides speed, accuracy, and agility that we never thought would be possible. Cost and the most expensive resource of all, human, has been a massive driver for us through our Alteryx journey and I'd expect that these drivers will continue as time passes us by.
Describe the working solution
Our processes vary from workflow to workflow, however overall we use a lot of SQL, Oracle, Teradata and SharePoint. In some workflows we blend 2 sources; in others we blend all of them. It depends on the need of the business that we are working with on any given day. Once the blending is done we do a variety of things with it, sometimes is goes to apps for self-service consumption and other times we push it into a data warehouse. However one thing that is consistent in our process is final data visualization in Tableau! Today, upwards of 95% of our workflows end up in Tableau allowing us to empower our users with self-service and analytics reporting. When using databases like SQL and Oracle we see MASSIVE gains in the use of In-Database tools. The ability for our Alteryx users to leverage such a strong no code solution creates an advantage for us in the customer service and analytics space because they already understand the data but now they have a means to get to it.
Audit Automation:
Billing:
File Generator:
Market Generator:
Parse:
Describe the benefits you have achieved
The 7,736 hours mentioned above is cumulative of 7 different processes that we rely on, on a regular basis.
We hope you've found our write-up compelling and win-worthy!
Author: Alex Huang, Asst. Mgr, Quality Planning & Analysis
Company: Hyundai Motor America
Awards Category: Most Time Saved
There have been just a few times where some tool or platform has truly "changed" my life. The two that come immediately to mind are Alteryx & Tableau. Before I had either, the majority of my time was spent wrangling data, creating reports, and doing what I could using SAS, SQL, & Excel. I had streamlined as much as I could and still felt bogged down by the rudimentary data tasks that plague many of us.
With the power of Alteryx alone, I've regained 1,253 hours per year. Alteryx WITH Tableau has saved me an additional 511 hours to a total of 1,764 hours saved per year! Does that mean I can quit? Maybe…but I’m not done yet!
For those that care for the details, here's a table of time savings I had cataloged during the start of my Alteryx journey. I’ve had to blank out the activity names for security reasons but the time savings are real.
I experienced a 71% savings in time with Alteryx alone!
With this new found "free time," I was able to prototype ideas stuck on my To-Do list and create new insight for my business unit. Now my "what if's" go from idea straight to Alteryx (and to Tableau faster) and I couldn't be happier. Insights are delivered faster than ever and with more frequent (daily) updates thanks to Alteryx Desktop Automation.
Describe the problem you needed to solve
Hyundai Motor America sells thousands of cars per day so the faster we can identify a quality issue and fix it, the more satisfied our customers will be. Addressing quality concerns earlier and faster helps us avoid additional costs but most importantly brand loyalty, perceived quality, and vehicle dependability, etc. Some examples of actions:
I needed to solve a time problem first! I was spending too much time doing things like data prep and reporting that just wasn’t quite enough for me. I didn't have enough time to do what I really wanted to do, solve problems!
Being an avid fan/user of Tableau, data preparation started becoming my biggest challenge as my dashboard library grew. I would end up writing monster SQL statements and scripts to get the data ready but I still struggled with automation for creating Tableau Data Extracts (TDE's). I explored using Python to create them but it just wasn't quite the "desired" experience. Enter Alteryx, life changed.
Describe the working solution
My work typically involves blending data from our transactional data warehouse, call center data, survey data, and blending third-party data from companies like J.D. Powers. Since we have an Oracle database in-house, I'm able to leverage the In-DB tools in Alteryx which is just amazing! In-DB tools are similar to a "visual query builder" but with the Alteryx look, feel, and added capability of Dynamic Input and Macro Inputs. Since data only moves out of the DB when you want it to, queries are lightning fast which enable accelerated prototyping ability!
Describe the benefits you have achieved
I've quite literally freed up 93% of my time (given 1960 work hours per year with 15 days of vacation @ 8 hours per day) and started a new "data team" within my business unit with Alteryx & Tableau at its core. The ultimate goal will be to replicate my time savings for everyone and “free the data” through self-service apps. At this point, I’ve deployed 5,774 Alteryx nodes using 61 unique tools in 76 workflows of which 24% or so are scheduled and running automatically. Phew! Props to the built-in “Batch Macro Module Example” for allowing me to calculate this easily!
We are able to identify customer pain points through an automated Alteryx workflow and algorithm that gauges how likely an issue will persist across all owners of the same model/trim package. We’ve seen how blending Experian ConsumerView data bolsters this model but we’re still in the cost justification phase for that. Upon detection of said pain point, we are able to trigger alerts and treatments across the wider population to mitigate the impact of this pain point. Issues that can’t be readily fixed per se are relayed back to R&D for further investigation. Ultimately customers may never see an issue because we’ve addressed it or they are simply delighted by how fast we’ve responded even when no immediate remedy is available.
The true bottom line is that the speed and accuracy at which we execute is critical in our business. Customers want to be heard and they want to know how we are going to help resolve their problems now, not months later. They want to love their Hyundai’s and the more they feel like we are helping them achieve that, the more loyal they will be to our brand.
Although we can’t fix everything, Alteryx helps us get to where we need to be faster which; in my opinion, is an enabler for success.
Author: Omid Madadi, Developer
Company: Southwest Airlines Co.
Awards Category: Best Business ROI
Describe the problem you needed to solve
Fuel consumption expense is a major challenge for the airline industry. According to the International Air Transport Association, fuel represented 27% of the total operating costs for major airlines in 2015. For this reason, most airlines attempt to improve their operational efficiency in order to stay competitive and increase revenue. One way to improve operational efficiency is to increase the accuracy of fuel consumption forecasting.
Currently, Southwest Airlines offers services in 97 destinations with an average of 3,500 flights a day. Not having enough fuel at an airport is extremely costly and may result in disrupting flights. Conversely, ordering more fuel than what an airport needs results in high inventory and storage costs. As such, the objective of this project was to develop proper forecasting models and methods for each of these 97 airports in order to increase the accuracy and speed of fuel consumption by using historical monthly consumption data.
Describe the working solution
Data utilized in this project were from historical Southwest Airlines monthly fuel consumption reports. Datasets were gathered from each of the 97 airports as well as various Southwest departments, such as finance and network planning. Forecasting was performed on four different categories: scheduled flights consumption, non-scheduled flights consumption, alternate fuel, and tankering fuel. Ultimately, the total consumption for each airport was obtained by aggregating these four categories. Since data were monthly, time series forecasting and statistical models - such as autoregressive integrated moving average (ARIMA), time series linear and non-linear regression, and exponential smoothing - were used to predict future consumptions based on previously observed consumptions. To select the best forecasting model, an algorithm was developed to compare various statistical model accuracies. This selects a statistical model that is best fit for each category and each airport. Ultimately, this model will be used every month by the Southwest Airlines Fuel Department.
In addition to developing a consumption forecast that increases fuel efficiency, a web application was also developed. This web application enables the Fuel Department to browse input data files, upload them, and then run the application in an easy, efficient, and effortless manner. Data visualization tools were also added to provide the Fuel Department with better insights of trends and seasonality. Development of the statistical models has been finalized and will be pushed to production for use by the Southwest Airlines Fuel Department soon.
Describe the benefits you have achieved
Initially, the forecasting process for all 97 Southwest Airlines airports used to be conducted through approximately 150 Excel spreadsheets. However, this was an extremely difficult, time-consuming, and disorganized process. Normally, consumption forecasts would take up to three days and would have to be performed manually. Furthermore, accuracy was unsatisfactory since Excel's capabilities are inadequate in terms of statistical and mathematical modeling.
For these reasons, a decision was made to use CRAN R and Alteryx for data processing and development of the forecasting models. Alteryx offers many benefits since it allows executing R language script by using R-Tool. Moreover, Alteryx makes data preparations, manipulations, processing, and analysis fast and efficient for large datasets. Multiple data sources and various data types have been used in the design workflow. Nonetheless, Alteryx made it convenient to select and filter input data, as well as join data from multiple tables and file types. In addition, the Fuel Department needed a web application that would allow multiple users to run the consumption forecast without the help of any developers, and Alteryx was a simple solution to the Fuel Department's needs since it developed an interface and published the design workflow to a web application (through the Southwest Airlines' gallery).
In general, the benefits of the consumption forecast include (but are not limited to) the following:
Author: Qin Lee, Business Analyst
Company: MSXI
Awards Category: Most Unexpected Insight
Huge data, large file and multiple applications have been created and saved and shared in a small size of Alteryx file. And now, I can test the script/coding and find the errors. This is the good way to develop the proof of concept for our company.
Describe the problem you needed to solve
We need to go through many applications to get the data and save into one location to share and view.
Describe the working solution
We are blending the data sources form SQL, Access Excel and Hadoop, Yes, we are leveraging many parties' data. We are developing the workflows and functions for a concept now. Yes, we are exporting to a visualization tool.
Describe the benefits you have achieved
Collected the data from many locations and saved into a small size of the Alteryx database file and created the workflow and function and developed a search engine and design the proof of concept for approval and launch. Saved time and resolved the problem and increased customer satisfaction. I would like to send my sincere thanks to Mr. Mark Frisch (@MarqueeCrew), who helped us for many days to finish this project.
Author: Shelley Browning, Data Analyst
Company: Intermountain Healthcare
Awards Category: Most Time Saved
Describe the problem you needed to solve
Intermountain Healthcare is a not-for-profit health system based in Salt Lake City, Utah, with 22 hospitals, a broad range of clinics and services, about 1,400 employed primary care and secondary care physicians at more than 185 clinics in the Intermountain Medical Group, and health insurance plans from SelectHealth. The entire system has over 30,000 employees. This project was proposed and completed by members of the Enterprise HR Employee Analytics team who provide analytic services to the various entities within the organization.
The initial goal was to create a data product utilizing data visualization software. The Workforce Optimization Dashboard and Scorecard is to be used throughout the organization by employees with direct reports. The dashboard provides a view of over 100 human resource metrics on activities related to attracting, engaging, and retaining employees at all levels of the organization. Some of the features in the dashboard include: drilldown to various levels of the organization, key performance indicators (KPI) to show change, options for various time periods, benchmark comparison with third party data, and links to additional resources such as detail reports. Prior to completion of this project, the data was available to limited users in at least 14 different reports and dashboards making it difficult and time consuming to get a complete view of workforce metrics.
During initial design and prototyping it was discovered that in order to meet the design requirements and maintain performance within the final visualization it would be necessary for all the data to be in a single data set. The data for human resources is stored in 17 different tables in an Oracle data warehouse. The benchmark data is provided by a third party. At the time of development the visualization software did not support UNION or UNION ALL in the custom SQL function. During development the iterative process of writing SQL, creating an extract file, and creating and modifying calculations in the visualization was very laborious. Much iteration was necessary to determine the correct format of data for the visualization.
Other challenges occurred, such as when it was discovered that the visualization software does not support dynamic field formatting. The data values are reported in formats of percent, currency, decimal and numeric all within the same data column. While the dashboard was in final review it was determined that a summary of the KPI indicators would be another useful visualization on the dashboard. The KPI indicators, red and green arrows, were using table calculations. It is not possible to create additional calculations based on the results of table calculations in the visualization software. The business users also requested another cross tabular view of the same data showing multiple time periods.
Describe the working solution
Alteryx was instrumental in the designing and development of the visualization for the workforce dashboard. Without Alteryx the time to complete this project would have easily doubled. By using Alteryx, a single analyst was able to iterate through design and development of both the data set and the dashboard.
The final dashboard includes both tabular and graphic visualizations all displayed from the same data set. The Alteryx workflow uses 19 individual Input Data tools to retrieve data from the 17 tables in Oracle and unions this data into the single data set. Excel spreadsheets are the source for joining the third party benchmark data to the existing data. The extract is output from Alteryx directly to a Tableau Server. By utilizing a single set of data, filtering and rendering in visualization are very performant on 11 million rows of data. (Development included testing data sets of over 100 million rows with acceptable but slower performance. The project was scaled back until such a time as Alteryx Server is available for use.)
Describe the benefits you have achieved
The initial reason for using Alteryx was the ability to perform a UNION ALL on the 19 input queries. By selecting the option to cache queries, output directly to tde files, and work iteratively to determine the best format for the data in order to meet design requirements and provide for the best performance for filtering and rendering in the visualization, months of development time was saved. The 19 data inputs contain over 7000 lines of SQL code combined. Storing this code in Alteryx provides for improved reproducibility and documentation. During the later stages of the project it was fairly straight forward to use the various tools in Alteryx to transform the data to support the additional request for a cross tab view and also to recreate the table calculations to mimic the calculations the visualization. Without Alteryx it would have taken a significant amount of time to recreate these calculations in SQL and re-write the initial input queries.
Our customers are now able to view their Workforce Optimization metrics in a single location. They can now visualize a scenario in which their premium pay has been increasing the last few pay periods and see that this may be attributed to higher turnover rates with longer times to fill for open positions, all within a single visualization. With just a few clicks our leaders can compare their workforce optimization metrics with other hospitals in our organization or against national benchmarks. Reporting this combination of metrics had not been attempted prior to this time and would not have been possible at this cost without the use of Alteryx.
Costs saving are estimated at $25,000 to-date with additional savings expected in future development and enhancements.
Author: Michael Barone, Data Scientist
Company: Paychex Inc.
Awards Category: Most Time Saved
We currently have more than two dozen predictive models, pulling data of all shapes and sizes from many different sources. Total processing time for a round of scoring takes 4 hours. Before Alteryx, we had a dozen models, and processing took around 96 hours. That's a 2x increase in our model portfolio, but a 24x decrease in processing time.
Describe the problem you needed to solve
Our Predictive Modeling group, which began in the early-to-mid 2000s, had grown from one person to four people by summer 2012. I was one of those four. Our Portfolio had grown from one model, to more than a dozen. We were what you might call a self-starting group. While we had the blessing of upper Management, we were small and independent, doing all research, development, and analysis ourselves. We started with using the typical every day Enterprise solutions for software. While those solutions worked well for a few years, by the time we were up to a dozen models, we had outgrown those solutions. A typical round of "model scoring" which we did at the beginning of ever y month, took about two-and-a-half weeks, and ninety-five percent of that was system processing time which consisted of cleansing, blending, and transforming the data from varying sources.
Describe the working solution
We blend data from our internal databases - everything from Excel and Access, to Oracle, SQL Server, and Netezza. Several models include data from 3rd party sources such as D&B, and the Experian CAPE file we get with out Alteryx data package.
Describe the benefits you have achieved
We recently have taken on projects that require us processing and analyzing billions of records of data. Thanks to Alteryx and more specifically the Calgary format, most of our time is spent analyzing the data, not pulling, blending, and processing. This leads to faster delivery time of results, and faster business insight.
Author: Andy Kriebel (@VizWizBI), Head Coach
Company: The Information Lab
Awards Category: Best 'Alteryx for Good' Story
The Connect2Help 211 team outlined their requirements, including review the database structure and what they were looking for as outputs. Note that this was also the week that we introduced Data School to Alteryx. We knew that the team could use Alteryx to prepare, cleanse and analyse the data. Ultimately, the team wanted to create a workflow in Alteryx that Connect2Help 211 could use in the future.
Ann Hartman, Director of Connect2Help 211 summarized the impact best: "We were absolutely blown away by your presentation today. This is proof that a small group of dedicated people working together can change an entire community. With the Alteryx workflow and Tableau workbooks you created, we can show the community what is needed where, and how people can help in their communities."
The entire details of the project can be best found here - http://www.thedataschool.co.uk/andy-kriebel/connect2help211/
Describe the problem you needed to solve
In July 2015, Connect2Help 211, an Indianapolis-based non-profit service that facilitates connections between people who need human services and those who provide them, reached out to the Tableau Zen Masters as part of a broader effort that the Zens participate in for the Tableau Foundation. Their goals and needs were simple: Create an ETL process that extracts Refer data, transforms it, and loads it into a MYSQL database that can be connected to Tableau.
Describe the working solution
See the workflow and further details in the blog post - http://www.thedataschool.co.uk/andy-kriebel/connect2help211/
Describe the benefits you have achieved
While the workflow looks amazingly complex, it absolutely accomplished the goal of creating a reusable ETL workflow. Ben Moss kicked off the project presentations by taking the Connect2Help 211 team through what the team had to do and how Connect2Help 211 could use this workflow going forward.
From there, the team went through the eight different visualisation that they created in Tableau. Keep in mind, Connect2Help 211 wasn't expecting any visualisations as part of the output, so to say they were excited with what the team created in just a week is a massive understatement.